[go: nahoru, domu]

US11701208B2 - Detecting tooth shade - Google Patents

Detecting tooth shade Download PDF

Info

Publication number
US11701208B2
US11701208B2 US16/946,186 US202016946186A US11701208B2 US 11701208 B2 US11701208 B2 US 11701208B2 US 202016946186 A US202016946186 A US 202016946186A US 11701208 B2 US11701208 B2 US 11701208B2
Authority
US
United States
Prior art keywords
tooth
shade
texture
digital
representation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US16/946,186
Other versions
US20200352688A1 (en
Inventor
Bo ESBECH
Rune Fisker
Lars Henriksen
Tais Clausen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
3Shape AS
Original Assignee
3Shape AS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=52473894&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=US11701208(B2) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Application filed by 3Shape AS filed Critical 3Shape AS
Priority to US16/946,186 priority Critical patent/US11701208B2/en
Assigned to 3SHAPE A/S reassignment 3SHAPE A/S ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CLAUSEN, TAIS, FISKER, RUNE, ESBECH, Bo, HENRIKSEN, LARS
Publication of US20200352688A1 publication Critical patent/US20200352688A1/en
Priority to US17/742,813 priority patent/US11723759B2/en
Priority to US17/742,955 priority patent/US11707347B2/en
Application granted granted Critical
Publication of US11701208B2 publication Critical patent/US11701208B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61CDENTISTRY; APPARATUS OR METHODS FOR ORAL OR DENTAL HYGIENE
    • A61C13/00Dental prostheses; Making same
    • A61C13/08Artificial teeth; Making same
    • A61C13/082Cosmetic aspects, e.g. inlays; Determination of the colour
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J3/00Spectrometry; Spectrophotometry; Monochromators; Measuring colours
    • G01J3/46Measurement of colour; Colour measuring devices, e.g. colorimeters
    • G01J3/50Measurement of colour; Colour measuring devices, e.g. colorimeters using electric radiation detectors
    • G01J3/504Goniometric colour measurements, for example measurements of metallic or flake based paints
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J3/00Spectrometry; Spectrophotometry; Monochromators; Measuring colours
    • G01J3/46Measurement of colour; Colour measuring devices, e.g. colorimeters
    • G01J3/50Measurement of colour; Colour measuring devices, e.g. colorimeters using electric radiation detectors
    • G01J3/508Measurement of colour; Colour measuring devices, e.g. colorimeters using electric radiation detectors measuring the colour of teeth
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H20/00ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance
    • G16H20/40ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to mechanical, radiation or invasive therapies, e.g. surgery, laser therapy, dialysis or acupuncture
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/50ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for simulation or modelling of medical disorders

Definitions

  • This invention generally relates to methods and a user interfaces for determining the shade of a patient's tooth or teeth and for utilizing the determined tooth shades for designing and manufacturing dental restorations.
  • both the shape and shade of the manufactured restoration is adapted to the patient's natural teeth surrounding the restoration. If the shade of the restoration differs significantly from the surrounding natural teeth, e.g. is significantly darker or brighter than these, the restoration appear artificial and deteriorate the aesthetic impression of the patient's smile.
  • the tooth color can be represented in many different color spaces, such as the L*C*h*color space representing color in terms of Lightness, Chroma and Hue, or in the L*a*b*color space as described e.g. by Hassel et al. (Hassel 2012) and Dozic et al. (Dozic 2007).
  • the L*a*b*color space has the advantage that it is designed to approximate human vision with the L*component closely matches human perception of lightness.
  • each reference tooth shade value in a tooth shade guide represents a predetermined and known tooth color value and often correspond to the color of commercially available ceramics for the production of dental restorations. This is e.g. the case for the VITA 3D-Master or the VITA Classic shade guides provided by VITA Zahnfabrik, Germany.
  • tooth shades are expressed in codes referring to the L*C*h* color space, where each code is constructed according to (Lightness, hue, Chroma).
  • a tooth shade value is 3R1.5 where “3” refers to the lightness, “R” to the hue and “1.5” to the Chroma of the tooth.
  • the dentist When manually determining which reference tooth shade value best matches the color of a patient's tooth, the dentist holds different pre-manufactured teeth of the shade guide at the tooth for comparison. Often a picture is taken with the pre-manufactured structures arranged at the teeth. The technician who produces the prosthetic then uses the picture in evaluating which ceramic must be used for the different parts of the restoration based on the picture. This process is both time consuming and inaccurate.
  • a method for determining shade of a patient's tooth comprising:
  • a user interface for determining and displaying shade of a patient's tooth, wherein the user interface is configured for:
  • the texture data of the digital 3D representation expresses the texture of the tooth.
  • the texture data can be a texture profile expressing the variation in the texture over the tooth.
  • the shape data of the digital 3D representation expresses the shape of the tooth.
  • the texture information comprises at least one of tooth color or surface roughness.
  • the texture data expressing a texture profile of the tooth may be color data expressing a color profile of the tooth, and the tooth shade value for a point on the tooth may be derived by comparing the color data of the corresponding point of the digital 3D representation with known color values of one or more reference tooth shade values.
  • Determining both the tooth shade value from a digital 3D representation comprising both shape data expressing the shape of the tooth and texture data expressing a texture profile of the tooth provides the advantage that shade and geometry information are directly linked.
  • CAD Computer Aided Design
  • CAM Computer Aided Design
  • the material used for the manufacture of the dental restoration can then be selected based on the determined tooth shade value.
  • the dental restoration is manufactured with a shade profile where the shade differs from the incisal edge towards cervical end of the restoration.
  • the disclosed invention allows the operator to determine tooth shade values for several points on the tooth such that a shade profile can be determined for the dental restoration.
  • Multi-shaded milling blocks exits which mimics standard tooth shade profiles. Having the shape data and the tooth shade values linked via the digital 3D representation provides that the correct portion of the multi-shaded milling block can be milled out. The remaining portion of the multi-shaded milling block forming the dental restoration will then have a shape and shade profile which closely resembles that of a natural tooth.
  • obtaining the digital 3D representation of the tooth comprises recording a series of sub-scans of the tooth, where at least one of said sub-scans comprises both texture information and geometry information for said tooth, and generating the digital 3D representation of the tooth from the recorded series of sub-scans.
  • the texture data for the digital 3D representation can be derived by combining the texture information of the several sub-scans.
  • the recorded sub-scans comprise at least data of the tooth for which the shade is determined, but potentially also of the neighboring teeth such that for example the shape and location of the neighboring teeth can be taken into account when designing a dental restoration for the tooth.
  • Texture information and texture data for the neighboring teeth can also be used to determine the shade value for the tooth, e.g. by interpolation of the shades determined for the neighbor teeth.
  • the method comprises creating a shade profile for the tooth from shade values determined for one or more of points on the tooth.
  • the shade profile of natural teeth often has a brighter shade at the incisal edge of the tooth and gradually changes into a darker shade towards the cervical end of the tooth, i.e. the end at the patient's gingiva.
  • the tooth shade profile may be generated based on knowledge of the normal tooth shade profile for that particular type of tooth and patient. This knowledge may relate to how the shade profile normally changes over the tooth, the age and gender of the patients, etc.
  • the profile will be based on tooth shades determined in several points on the tooth to provide the most reliable tooth shade profile.
  • the user interface is configured for creating a shade profile for the tooth from tooth shade values determined for one or more points on the tooth.
  • the tooth shade profile can be created by interpolation of tooth shade values determined for points distributed over the tooth surface with some distance between the points.
  • the tooth shade value for some parts of the tooth surface are then not derived directly from sub-scan texture information relating to these parts but from the determined tooth shade values for other parts/points on the tooth surface.
  • a tooth shade profile for the entire labial/buccal surface of the tooth can thus be created from a selection of points on the surface providing a fast and often sufficiently accurate procedure for creating the tooth shade profile.
  • the interpolation of the tooth shade values can be realized by an interpolation in each of the coordinates of the color space used to describe the tooth color.
  • the tooth shade profile comprises a one or more tooth shade regions on the tooth surface where an average tooth shade is derived for each region from tooth shade values determined for a number of points within the region.
  • the tooth shade region can be defined by a structure encircling a portion of the tooth surface in the digital 3D representation, where either the operator or a computer implemented algorithm decides where each geometric structure is located on the digital 3D representation.
  • Different shapes e.g. circles, squares, or rectangles
  • sizes e.g. corresponding to a few millimeters
  • the number of points within the geometrical structure can be increased to provide a more accurate measure of the shade or reduced to provide a faster calculation.
  • the average tooth shade value for a region can e.g. be derived as a weighted average where the tooth shade value for points in the center of the structure is assigned a higher weight than tooth shade value of points closer to the boundary.
  • the tooth surface can also be divided into a coronal, a middle and a cervical region.
  • Some natural teeth has a shade profile which can be expressed by such a division and many dentists and dental technicians are familiar with such a division.
  • tooth shade values are determined for a plurality of teeth, i.e. on parts of the digital 3D representation corresponding to two or more teeth, and a tooth shade value and/or a tooth shade profile for each of these teeth is created from the determined tooth shade values.
  • the texture data at least partly are derived by combining the texture information from corresponding parts of a number of the sub-scans.
  • the digital 3D representation can be generated through registration of sub-scans into a common coordinate system by matching overlapping sections of sub-scans, i.e. the sections of the sub-scans which relate to the same region of the tooth.
  • deriving the texture data for this region in the digital 3D representation can comprise combining the corresponding texture information, i.e. the texture information in the sub-scans corresponding to the same sections of the tooth.
  • Deriving the texture data based on texture information from two or more sub-scan can provide a more accurate measurement of the texture data.
  • the texture information of one sub-scan for a particular region of the tooth may be unreliable e.g. due to the angle between the surface in this region and the scanner when this particular sub-scan was recorded.
  • the combination of texture information from several sub-scans can provide a more reliable color.
  • combining the texture information from the sub-scans comprises interpolating the texture information, i.e. texture information from parts of the sub-scans corresponding to a point on the tooth are interpolated to determine the texture data for that point.
  • Such an interpolation can provide that the determined texture data is more accurate e.g. in cases where the texture information for a point on the tooth is not linearly varying over the sub-scans such that a simple averaging will not provide the best result.
  • combining the texture information from the sub-scans comprises calculating an average value of the texture information, i.e. texture data for a point on the digital 3D representation are determined by averaging the texture information of the sub-scans corresponding to that point on the tooth.
  • the calculated average value is a weighted average of the texture information.
  • This approach has the advantage that the derived texture data of the digital 3D representation are not as sensitive to errors in the texture information of a single sub-scan.
  • the texture data for a point on the tooth is preferably derived from a number of sub-scans where at least some of the sub-scans are recorded at different orientations of the scanner relative to the teeth. The sections of the sub-scans relating to this point are hence acquired at different angles relative to the tooth surface in this point.
  • a portion of a sub-scan recorded from a surface perpendicular to the optical path of the probe light at the tooth may be dominated by specular reflected light which does not describe the texture of the tooth but rather the spectral distribution of the probe light.
  • a portion of a sub-scan recorded from a tooth surface almost parallel to the optical path is often quite weak and hence often provide an erroneous detection of the texture at that point.
  • the texture information from parts of a sub-scan relating to a tooth surface which is substantially perpendicular or parallel to the optical path are assigned a low weight in the weighted averaging of the texture information to determine the texture data for the point.
  • the orientation of the scanner relative to the tooth when a sub-scan is acquired can be determined from the shape of the sub-scan. Parts of the sub-scan relating to tooth surfaces which are substantially parallel or perpendicular to the optical path can thus immediately be detected in the sub-scan such that the texture information of the corresponding parts are assigned at low weight when determining the texture data for this point from a series of sub-scans.
  • a specular reflection from the tooth often has an intensity which is significantly higher than that of e.g. diffuse light from surfaces which have an oblique angle relative to the optical path. In some cases the specular reflection will saturate the pixels of the image sensor used for the recording of the sub-scans.
  • the method comprises detecting saturated pixels in the recorded sub-scans and assigning a low weight to the texture information of the saturated pixels when combining the texture information from the sub-scans, i.e. when calculating the weighted average of the texture information.
  • Specular reflection from a tooth surface may also be detected from a comparison between the spectrum of the light received from the tooth and that of the probe light. If these spectra a very similar it indicates that the tooth has a perfectly white surface which is not natural. Such texture information may thus be assigned a low weight in a weighted average of texture information.
  • determining the tooth shade value for the point comprises selecting the reference tooth shade value with known texture value closest to the texture data of the point.
  • selecting the tooth shade value of the point can comprise calculating the color difference between the determined color data in the point and the color data of the reference tooth shade values.
  • This difference can e.g. be calculated as a Euclidian distance in the used color space.
  • Dozic et al. describes that the Euclidian distance ⁇ E between two points (L* 1 , a* 1 , b* 1 ) and (L* 2 , a* 2 , b* 2 ) in the L*a*b*color space is given by:
  • Selecting the tooth shade value can then comprise determining for which of the reference tooth shades the color difference, i.e. the Euclidian distance, is the smallest.
  • determining the tooth shade value for the point comprises an interpolation of the two or more reference tooth shade values having known texture values close to the texture data of the point.
  • tooth shade can be represented with a more detailed solution than what is provided by the tooth shade standard used to describe the tooth shade. For instance when using a Lightness-Hue-Chroma code a tooth shade value of 1.5M2.5 can be determined for the tooth by interpolation of Lightness values of 1 and 2, and Chroma values of 2 and 3.
  • the tooth shade value can be displayed in a user interface e.g. together with the digital 3D representation of the tooth. If the digital 3D representation also contains parts relating to other teeth the tooth shade value for the tooth is preferably displayed at the tooth, such as at the point of the for which the tooth shade value has been determined.
  • the tooth shade value can also be represented as a color mapped onto the digital 3D representation.
  • the method comprises deriving a certainty score expressing the certainty of the determined tooth shade value.
  • Deriving a certainty score for the determined tooth shade value provides the advantage that a measure of how accurate the determined value is can be displayed to the operator, preferably when the patient is still at the clinic such that further scanning can be performed if this is required to provide a more precise tooth shade value.
  • the method comprises generating a visual representation of the certainty score and displaying this visual representation in a user interface.
  • the method comprises generating a certainty score profile at least for a portion of the tooth, where the certainty scope profile represents the certainty scores for tooth shade values determined for a number of points on the tooth, such as for the values in a tooth shade profile for the tooth.
  • the certainty score profile can be mapped onto the digital 3D representation of the tooth and visualized in a user interface. When the tooth shade profile also is mapped onto the tooth digital 3D representation the operator may be allowed to toggle between having the tooth shade profile and having the certainty scope profiled visualized on the digital 3D representation.
  • the visual representation of the certainty score is displayed together with or is mapped onto the digital 3D representation of the tooth.
  • the method comprises comparing the derived certainty score with a range of acceptable certainty score values. This is done to verify that the certainty score is acceptable, i.e. that the determined tooth shade value is sufficiently reliable.
  • One boundary of the range can be defined by a threshold value.
  • the threshold value may define the lower boundary of the range and vice versa.
  • a visual representation of the certainty score or of the result of the comparison of the certainty score with the range can be generated and displayed in a user interface. Preferably, this visual representation is displayed together with the determined tooth shade value.
  • the method comprises deciding based on the certainty score whether the determined tooth shade value or tooth shade profile is acceptable. This may be based on the comparison of the derived certainty score and the range of acceptable certainty score values, e.g. where it is decided that the determined tooth shade value is acceptable if the certainty score is within the range of acceptable values.
  • the certainty measure relates to how uniform the sub-scan texture information is at the point.
  • the texture data derived therefrom may be unreliable and the tooth shade value derived for this point is accordingly not very reliable.
  • the certainty measure relates to how close the texture data is to the known texture value of the determined tooth shade value.
  • the certainty measure may relate to how close one parameter of the color data of the digital 3D representation is to the corresponding parameter of the known color for the determined tooth shade value.
  • the certainty measure may relate to the difference in the lightness parameter between point of the digital 3D representation and the determined tooth shade value.
  • the Euclidian distance between the color data to the selected reference tooth shade value can also be used in determining the certainty measure. If the Euclidian distance is above a threshold value the uncertainty is then evaluated to be too large.
  • the color data can here both relate to color data of the point or the average color data for a region surrounding the point.
  • the certainty measure relates to the amount of texture information used to derive the texture data at the point.
  • the texture data for the point is derived from a limited amount of texture information the texture data, and accordingly the tooth shade value derived therefrom, may be less reliable than the tooth shade values derived from large amounts of texture information.
  • the visual representation of the certainty score comprises a binary code, such as red for certainty scores outside a range of acceptable certainty score values, and green for certainty scores within the range, a bar structure with a color gradient, a numerical value, and/or a comparison between the texture data and the known texture value of the determined tooth shade value.
  • the visual representation of the certainty score comprises a certainty score indicator.
  • the certainty score indicator may comprise a bar structure with a color gradient going from a first color representing a low certainty score to a second color representing a high certainty score.
  • the first color may be red and the second color green.
  • the color gradient of the bar structure may be configured to have an intermediate color, e.g. yellow representing the threshold value for the certainty score.
  • the certainty score indicator may comprise marker which is arranged relative to the color gradient of the bar structure such that it indicated the certainty score.
  • the visual representation of the certainty score comprises a numerical value, such as a numerical value in an interval extending from a lower limit indicating a low certainly, i.e. a relatively uncertain tooth shade value, to a higher limit indicating a high certainty, i.e. a relatively certain tooth shade value.
  • the one or more reference tooth shade values relate to shade values for natural teeth with intact surface and/or to shade values for teeth prepared for a dental restoration.
  • the reference tooth shade values used for determining the tooth shade can be selected based on the tooth.
  • Intact and healthy teeth normally have tooth shades in one range of tooth shade values where a tooth prepared for a dental restoration has a tooth shade in another range, which may overlap with the range for healthy teeth. It may thus be advantageous that the operator enters whether the tooth is intact or prepared for a restoration and the appropriate color space is used in the comparison with the texture data.
  • the point may e.g. be on the gingiva of the patient or relate to silver filling.
  • the method comprises comparing the texture data with known texture values for soft oral tissue, such as gum tissue and gingiva.
  • This may e.g. be relevant when the certainty scores are outside said range of acceptable certainty score values for all tooth shade values of a tooth shade system, i.e. if there is a poor match between the texture data and the known texture for all the tooth shades of the reference set.
  • a user interface for implementing the method it may be suggested to the operator that the point perhaps is not on a tooth surface but on the gums or gingiva of the patient. This suggestion may be provided both when the texture data has been found to give a good match with known texture values of gum/gingiva and/or when the texture data has a poor match with the known texture values of the reference tooth shade values in the tooth shade system or systems.
  • the method comprises determining an alternative tooth shade value for the point when said certainty score is outside said range of acceptable certainty score values.
  • the method comprises displaying the alternative tooth shade value in the user interface optionally together with the digital 3D representation of the patient's set of teeth and/or the initially determined tooth shade value.
  • the digital 3D representation of the tooth is generated at least partly from the geometry information of the sub-scans.
  • the texture information of the sub-scans is also taken into account when generating the digital 3D representation of the tooth.
  • Sub-scans comprising texture information and geometry information may be recorded for more than said tooth, such that the generated digital 3D representation may comprise shade data expressing the shape and texture data expressing the texture profile of several of the patient's teeth.
  • a method for determining shade of a patient's tooth comprising:
  • a user interface for determining and displaying shade of a patient's tooth, wherein the user interface is configured for:
  • the user interface is configured for deriving a certainty score expressing the certainty of the determined tooth shade value for said point.
  • the user interface comprises a virtual tool which when activated on a point of the digital 3D representation of the tooth provides that
  • the user interface can then provide the operator with an opportunity to decide based on the visualized certainty score and/or the visual representations whether the determined tooth shade value or tooth shade profile is acceptable.
  • the visual representation of the comparison of the derived certainty score with the range of acceptable certainty score values comprises a binary code, such as red for certainty scores outside a range of acceptable certainty score values, and green for certainty scores within the range. Other means for this visualization are described above.
  • the visualized certainty score and/or the representation(s) of the certainty score or comparison of the certainty score with the range of acceptable certainty score values may be displayed at the digital 3D representation in the user interface or in a shade value region of the user interface.
  • the user interface is configured for determining an alternative shade value for the point and for displaying the alternative shade value when the certainty scores outside a range of acceptable certainty score values.
  • the digital restoration design can e.g. be for the manufacture of dental prosthetic restoration for the patient, such as a crown or a bridge restoration, where the digital restoration design expresses a desired shape and shade profile of the dental restoration.
  • Such digital restoration designs can be in the form of a CAD model of the dental restoration.
  • the method comprises suggesting a dental material for manufacturing the dental restoration from the digital restoration design based on the determined restoration shade.
  • the tooth shade value or tooth shade profile can be determined for the existing tooth and the shade of the digital restoration design based on the tooth shade value or tooth shade profile of the existing tooth.
  • This may e.g. be advantageous for the crown portions of a bridge restoration in the case where the tooth which is intended to accept the crown portion of the bridge is a healthy tooth.
  • the dental restoration is designed and manufactured for a tooth which either is damaged or has an undesired shade profile, such as for a broken or dead tooth.
  • the desired texture profile is derived by interpolation or averaging of the texture data of the digital 3D representation of the neighbor teeth.
  • one or more of the sub-scans comprise texture information for the patient's soft tissue, and optionally geometry information for said soft tissue.
  • the generated digital 3D representation may then comprise shape data expressing the shape of the soft tissue and texture data expressing a texture profile of the soft tissue.
  • an aesthetical pleasing denture can be designed where the color of the soft tissue part of the denture is selected based on the texture profile of the corresponding part of the digital 3D representation.
  • Knowledge of the texture of the soft tissue can also be used for diagnostics.
  • a warning may be prompted in a user interface to alert the operator that the soft tissue is suspicious.
  • a system for determining shade of a patient's tooth wherein the system comprises:
  • the sub-scans are recorded using an intra-oral scanner, such as the 3Shape TRIOS intra-oral scanner.
  • the intra-oral scanner may be configured for utilizing focus scanning, where the sub-scans of the scanned teeth are reconstructed from in-focus images acquired at different focus depths.
  • the focus scanning technique can be performed by generating a probe light and transmitting this probe light towards the set of teeth such that at least a part of the set of teeth is illuminated.
  • Light returning from the set of teeth is transmitted towards a camera and imaged onto an image sensor in the camera by means of an optical system, where the image sensor/camera comprises an array of sensor elements.
  • the position of the focus plane on/relative to the set of teeth is varied by means of focusing optics while images are obtained from/by means of said array of sensor elements. Based on the images, the in-focus position(s) of each of a plurality of the sensor elements or each of a plurality of groups of the sensor elements may be determined for a sequence of focus plane positions.
  • the in-focus position can e.g. be calculated by determining the maximum of a correlation measure for each of a plurality of the sensor elements or each of a plurality of groups of the sensor elements for a range of focus planes as described in WO2010145669. From the in-focus positions, sub-scans of the set of teeth can be derived with geometry information relating to the shape of the scanned surface.
  • the image sensor is a color sensor and the light source provides a multispectral signal a plurality of the sub-scans can include both geometry information and texture information, such as color information, for said tooth.
  • a digital 3D representation of the set of teeth can then be generated from the recorded sub-scans by e.g. the use of an Iterative Closest Point (ICP) algorithm.
  • Iterative Closest Point is an algorithm employed to minimize the difference between two clouds of points. ICP can be used to reconstruct 2D or 3D surfaces from different scans or sub-scans.
  • the algorithm is conceptually simple and is commonly used in real-time. It iteratively revises the transformation, i.e. translation and rotation, needed to minimize the distance between the points of two raw scans or sub-scans.
  • the inputs are: points from two raw scans or sub-scans, initial estimation of the transformation, criteria for stopping the iteration.
  • the output is: refined transformation. Essentially the algorithm steps are:
  • the generated digital 3D representation formed by such a procedure comprises shape data expressing the shape of the tooth.
  • the texture information of the sub-scans can be used in various ways to provide that the generated digital 3D representation also comprises texture data expressing a texture profile of the tooth.
  • the part of the sub-scan relating to the same point on the tooth can be identified, e.g. during the ICP procedure.
  • the corresponding texture information of these parts of the sub-scans can then be combined to provide the texture data for that point.
  • the invention relates to a computer program product comprising program code means for causing a data processing system to perform the method according to any of the embodiments, when said program code means are executed on the data processing system, and a computer program product, comprising a computer-readable medium having stored there on the program code means.
  • the present invention relates to different aspects including the method and user interface described above and in the following, and corresponding methods and user interface, each yielding one or more of the described advantage, and each having one or more embodiments corresponding to the embodiments described above and/or disclosed in the appended claims.
  • FIG. 1 shows an example of a flow chart for an embodiment.
  • FIGS. 2 to 4 show parts of screen shots of user interfaces.
  • FIG. 5 shows steps of a method for designing a dental restoration.
  • FIG. 6 shows a schematic of a system for determining tooth shade values.
  • FIGS. 7 A- 7 D and 8 A- 8 B show schematics of intra-oral scanning.
  • FIGS. 9 A- 9 B illustrates one way of determining tooth shade values from texture data.
  • FIG. 1 shows an example of a flow chart 100 for an embodiment of the method for determining shade of a patient's tooth.
  • step 102 a series of sub-scans of the patient's set of teeth is recorded, where a plurality of said sub-scans comprises both texture information and shape information for the tooth.
  • a digital 3D representation of the tooth is generated from said sub-scans, where the digital 3D representation comprises texture data expressing a texture profile of the tooth.
  • the digital 3D representation further comprises shape data expressing the shape of the tooth such that the shape of the tooth can be visualized in a user interface.
  • a tooth shade value for a point on the tooth is determined based on the texture data. This is done at least in part by comparing the texture data of the corresponding point of the digital 3D representation with a known texture value of one or more reference tooth shade values.
  • the reference tooth shade values may be provided in the form of a library file and comprise tooth shade values and corresponding texture values based on e.g. the VITA 3D-Master and/or the VITA Classic tooth shade systems.
  • FIGS. 2 to 4 show parts of screen shots from user interfaces in which derived tooth shade values and visual representation of the corresponding certainty scores for a number of tooth regions are displayed at the digital 3D representations of the patient's set of teeth.
  • the point or points on the tooth for which the tooth shade value(s) is/are determined can be selected by an operator. This can be the case e.g. when the digital 3D representation of the tooth is visualized in a user interface and the operator uses a pointing tool, such as a computer mouse, to indicate where on the digital 3D representation of the tooth, he wishes to determine the tooth shade value.
  • the point or points can also be selected by a computer implemented algorithm based on predetermined positions on the digital 3D representation of the tooth, such as a point arranged at a certain distance to the incisal edge of the tooth.
  • the screen shot 210 seen in FIG. 2 shows three regions 212 , 213 , 214 on the digital 3D representation of the patient's set of teeth. Two of these 212 , 213 , are selected at the part of the digital 3D representation corresponding to the tooth 211 while the third 214 is selected on the soft tissue part 215 of the digital 3D representation.
  • Average tooth shade value for a region can be calculated by averaging over tooth shade values derived for a number of points within the region or by calculating an average texture value for the region and determining the average tooth shade value therefrom.
  • the average tooth shade values are displayed in tooth value sections 217 , 218 , 219 linked to the regions in the user interface.
  • tooth value sections 217 , 218 relating to the regions 212 , 213 two tooth shade values are displayed where the upper shade value is derived using known texture values corresponding to the reference tooth shade values of the VITA 3D-Master tooth shade system and the lower tooth shade values relates to the VITA Classic tooth shade system. It is also seen that for the region 213 closest to the gingiva, the tooth shade is determined to be 2L1.5 in the VITA 3D-Master system and B1 in the VITA Classic system.
  • the certainty scores for the derived tooth shade values are visualized as a certainty score indicator displayed next to the tooth shade values.
  • the visualization of the certainty score indicator is in the form of a checkmark which indicates that the certainty score is sufficiently good to provide that the derived tooth shade values can be relied upon.
  • the color of the checkmark may provide further information to the certainty score, such as in cases where a green checkmark indicates a more certain tooth shade value than a yellow checkmark.
  • the third region 214 is located at the patient's soft tissue. An anatomical correct tooth shade value can hence not be calculated from the texture data of that part of the digital 3D representation of the patient's teeth and the corresponding certainty scope is accordingly very low.
  • the visualization of the certainty score in the tooth value section 219 is hence a cross indicating that the derived shade value was rejected. Further no shade value is indicated in the tooth value section 219 .
  • the screen shot 310 seen in FIG. 3 shows two regions 312 , 314 on the digital 3D representation of the patient's set of teeth. One of these regions 312 is selected at the part of the digital 3D representation corresponding to the tooth 311 while the second region 314 is selected on the soft tissue part 315 of the digital 3D representation. Average tooth shade value for a region can be calculated as described above in relation to FIG. 2 . Shade value sections 317 , 319 are also displayed for the regions 312 , 314 .
  • Two tooth shade values 321 are derived for the region 312 and displayed in the corresponding tooth value section 317 , where the upper value is derived using known texture values corresponding to the reference tooth shade values of the VITA 3D-Master tooth shade system (derived tooth shade value is 1.5M1) and the lower value using the VITA Classic tooth shade system (derived tooth shade value is B1).
  • the certainty score is visualized in the form of a certainty score indicator 322 comprising a vertical bar with a color gradient going from red representing a poor certainty score to green representing a good certainty score.
  • the certainty score indicator has a marker indicating the certainty score on the bar.
  • tooth shade value 1.5M1 of the VITA 3D-Master system is more certain than the tooth shade value B1 of the VITA Classic system for this region.
  • the tooth shade value of 1.5M1 is found by interpolation of the reference tooth shades 1M1 and 2M2.
  • the second region 314 is located at the patient's soft tissue.
  • An anatomical correct tooth shade value can hence not be calculated from the texture data of that part of the digital 3D representation of the patient's teeth and the corresponding certainty scope is accordingly very low as seen in the vertical bars of tooth value section 319 .
  • FIG. 4 shows a screen shot 410 where determined tooth shade values are derived for a total of 15 regions on the digital 3D representation of the tooth 411 .
  • the tooth shade values are all derived based on the known texture values of the reference tooth shade values of the VITA 3D-Master tooth shade system.
  • the certainty scores are visualized in the form of a certainty score indicator comprising a vertical bar with a color gradient going from red representing a poor certainty score to green representing a good certainty score.
  • the certainty score for the region 412 is almost at maximum while the certainty score of the region 413 is much close to a threshold for acceptable certainty score values.
  • the points may be arranged in a grid over the part of the digital 3D representation of the tooth.
  • FIG. 5 shows steps of a method for designing a dental restoration.
  • a digital restoration design is created e.g. based on the shape data of a digital 3D representation of the patient's set of teeth and/or on template digital restoration design loaded from a library.
  • Template digital restoration designs may e.g. be used when the tooth is broken.
  • step 532 the tooth shade values of different points or regions of the teeth are derived from the texture data of the digital 3D representation of the patient's set of teeth.
  • a desired shade profile for the dental restoration can be determined. This can be based on e.g. feature extraction where shade values are extracted from the other teeth by e.g. identifying shade zones on these teeth and copying these zones to the dental restoration. It can also be based on established shade rules for teeth, e.g. a rule describing a relation between the tooth shades values or profiles of the canines and the anterior teeth.
  • step 533 the desired tooth shade value(s) for the dental restoration is merged into the digital restoration design.
  • a CAD model of the milling block is provided, where the CAD model comprises information of the shade profile of the milling block material.
  • the optimal position of the digital restoration design relative to the CAD model of the milling block is then determined in 535 , where different criteria can be apply to provide the best fit between the desired shade profile and what actually can be obtained as dictated by the shade profile of the milling block.
  • step 536 the dental restoration is manufactured from the milling block by removing milling block material until the dental restoration is shaped according to the digital restoration design.
  • FIG. 6 shows a schematic of a system for determining tooth shade values.
  • the system 640 comprises a computer device 642 comprising a computer readable medium 643 and a processor 644 .
  • the system further comprises a visual display unit 647 , a computer keyboard 645 and a computer mouse 646 for entering data and activating virtual buttons in a user interface visualized on the visual display unit 647 .
  • the visual display unit can be a computer screen.
  • the computer device 642 is capable of receiving a digital 3D representation of the patient's set of teeth from a scanning device 641 , such as the TRIOS intra-oral color scanner manufactured by 3 shape A/S, or capable of receiving scan data from such a scanning device and forming a digital 3D representation of the patient's set of teeth based on such scan data.
  • the obtained digital 3D representation can be stored in the computer readable medium 643 and provided to the processor 644 .
  • the processor is configured for implementing the method according to any of the embodiments. This may involve presenting one or more options to the operator, such as where to derive the tooth shade value and whether to accept a derived tooth shade value. The options can be presented in the user interface visualized on the visual display unit 647 .
  • RGB color space For instance a focus scanner can record series of 2D color images for the generation of sub-scans, where the color information is provided in the RGB color space.
  • the processor 644 then comprises algorithms for transforming the recorded color data into e.g. the L*a*b or L*C*h color spaces.
  • the system may further comprise a unit 648 for transmitting a digital restoration design and a CAD model of a milling block to e.g. a computer aided manufacturing (CAM) device 649 for manufacturing a shaded dental restoration or to another computer system e.g. located at a milling center where the dental restoration is manufactured.
  • the unit for transmitting the digital restoration design can be a wired or a wireless connection.
  • the scanning of the patient's set of teeth using the scanning device 641 can be performed at a dentist while deriving the tooth shade values can be performed at a dental laboratory.
  • the digital 3D representation of the patient's set of teeth can be provided via an internet connection between the dentist and the dental laboratory.
  • FIGS. 7 A- 7 D and 8 A- 8 B show schematics of intra-oral scanning.
  • Different scanner configurations can be used to acquire sub-scans comprising both shape and texture information.
  • the scanner is mounted on axes with encoders which provides that the sub-scans acquired from different orientations can be combined using position and orientation readings from the encoders.
  • the scanner operates by the focus-scanning technique the individual sub-scans of the tooth are derived from a sequence of 2D images obtained while scanning a focus plane over a portion of the tooth.
  • the focus scanning technique is described in detail in WO2010145669.
  • the shape information of the sub-scans for an object, such as a tooth can be combined by algorithms for stitching and registration as widely known in the literature.
  • Texture data relating to the tooth color can be obtained using a scanner having a multi-chromatic light source, e.g. a white light source and a color image sensor. Color information from multiple sub-scans can be interpolated and averaged by methods such as texture weaving, or by simply averaging corresponding color components of the sub-scans corresponding to the same point/location on the surface. Texture weaving is described by e.g. Callieri M, Cignoni P, Scopigno R. “Reconstructing textured meshes from multiple range rgb maps”. VMV 2002, Er Weg, Nov. 20-22, 2002.
  • the scanner 741 (here represented by a cross-sectional view of the scanner tip) is held in one position relative to the teeth 711 , 760 (also represented by a cross-sectional view) while recording a sequence of 2D images for one sub-scan.
  • the illustrated teeth can e.g. be the anterior teeth in the lower jaw.
  • the size of the Field of View (here represented by the full line 761 on the teeth) of the scanner is determined by the light source, the optical components and the image sensor of the scanner.
  • the Field of View 761 covers part of the surface of the tooth 711 and part of the surface of the neighbor tooth 760 .
  • the generated digital 3D representation can thus also contain data for the neighbor teeth. This is often advantageous, e.g.
  • the scanner is arranged such that the acquired sub-scan comprises shape and color information for the incisal edge 762 of the teeth.
  • the probe light rays 763 from the scanner corresponding to the perimeter of the Field of View are also shown in the Figure. These probe light rays 763 define the optical path 764 of the scanner probe light at the tooth 711 .
  • a digital 3D representation of the tooth can be generated by combining sub-scans acquired from different orientations relative to the teeth, e.g. by sub-scan registration.
  • Sub-scans acquired from three such different orientations are illustrated in FIGS. 7 B, 7 C and 7 D , where only the optical path 763 of the scanner probe light is used to represent the relative scanner/tooth orientation in FIGS. 7 C and 7 D .
  • the sub-scans (here represented by the full line 765 on the teeth) covers different but overlapping sections of the tooth surface such that the sub-scans can be combined by registration into a common coordinate system using e.g. an Iterative Closest Point (ICP) algorithm as described above.
  • ICP Iterative Closest Point
  • a segment of each of the sub-scans corresponds to the point P on the tooth surface.
  • the sub-scans are registered to generate a digital 3D representation of the tooth, a correlation between these segments is established and the texture information of these sub-scan segments can be combined to determine the texture data for point P on the generated digital 3D representation of the tooth.
  • the color data of the digital 3D representation can be derived by averaging over each of the L*, a*, and b*parameters of the sub-scans. For example, the L*parameter of the color data for a given point P is then given by
  • L * ( P ) 1 N ⁇ ⁇ i N ⁇ L i * ⁇ ( P )
  • N is the number of sub-scans used in deriving the texture data
  • L* i (P) is the L*parameter of the i'th sub-scan for the segment relating to P.
  • Equivalent expressions are true for the a*and b*parameters for point P.
  • the color parameters for each point on the digital 3D representation of the tooth can be determined for sections of or the entire surface of the tooth, such that the generated digital 3D representation comprises both shape and texture information about the tooth.
  • the spatial resolution of the color data does not necessarily have to be identical to the resolution of the shape data of the digital 3D representation.
  • the point P can be described e.g. in Cartesian, cylindrical or polar coordinates.
  • the tooth shade value for that point can be determined by comparing the derived color data with the known color data of the reference tooth shade values of a tooth shade guide such as the VITA 3D-Master.
  • FIG. 8 A- 8 B illustrates some potentially problematic tooth surface areas for particular arrangements of the scanner 841 relative to the tooth 811 .
  • FIG. 8 A shows two points P i and P ii on the tooth 811 where the tooth surface is either substantially perpendicular or parallel to the optical path, such that the texture information recorded at P i and P ii may be unreliable.
  • the tooth surface at P i is perpendicular to the optical path 864 i at point P i which introduces the risk of having specular reflections of the probe light.
  • the optical path 864 ii at point P ii is parallel to the tooth surface at P ii such that the signal recorded from this part of the tooth surface in this sub-scan is relatively weak. This may cause that the color information in this section of the sub-scan are unreliable.
  • the averaging of the color information described above in relation to FIG. 7 can be a weighted averaging where the color information of unreliable sub-scans segments are assigned a lower weight than others.
  • FIG. 8 B is indicated three different optical paths 864 i , 864 ii and 864 iii at which sub-scans are acquired.
  • the color information of the segments of the sub-scans recorded with optical paths 864 i and 864 ii should be given a lower weight that the color information of the segment of the sub-scan recorded with the optical path 864 iii.
  • ⁇ i (P) is the weight factor for the color information of the i'th sub-scan in the segment at P.
  • FIG. 9 A- 9 B illustrates how a tooth shade value for a point P on a tooth can be determined based on reference tooth shade values.
  • the color data (L* P , a* P , b* P ) has been determined, e.g. by combining the color information of a series of sub-scans used for generating the digital 3D representation. If the color information originally is recorded using the RGB color space it is transformed into the L*a*b*color space using algorithms known to the skilled person.
  • the color data of the digital 3D representation and the known color values of the reference tooth shades are expressed in the L*a*b*color space, and the reference tooth shades are those from the VITA classical shade guide.
  • the reference shade values of the Vita classical shade guide are: B1, A1, B2, D2, A2, C1, C2, D4, A3, D3, B3, A3.5, B4, C3, A4, and C4.
  • the color data of these reference shades can be provided by scanning the corresponding pre-manufactured teeth of the shade guide. These color data are then also initially obtained in the RGB color space and can be converted to the L*a*b color space using the same algorithms applied to the color information/data for the point P.
  • the tooth shade value for the point is determined as the reference tooth shade value which has the smallest Euclidian distance to the point in the L*a*b color space.
  • the Euclidian distance ⁇ E P-R 1 from the color (L* P , a* P , b* P ) to the known colors of the reference tooth shade values are calculated using the expression:
  • FIG. 9 A only the known colors (L* R1 , a* R1 , b* R1 ) and (L* R2 , a* R2 , b* R2 ) for the two closest reference values R 1 and R 2 , respectively, are illustrated for simplicity. It can be seen that the Euclidian distance in the color space from P to R 2 is the smallest, and the tooth shade in point P is hence selected as that of R 2 .
  • the certainty score for the tooth shade value determined for point P depends on how close the color data of the point P is to the known color value of the selected reference tooth shade value. This can be quantified by the Euclidian distance and since point P is not particularly close to R 2 in FIG. 9 A the determined tooth shade has a poor certainty value.
  • An alternative approach to using the Euclidian distance is to determine individual parameters of the tooth shade value one at a time. This approach can be used e.g. when the reference tooth shades values are those of the Vita 3D-master system.
  • the reference tooth shade values of the Vita 3D-master shade guide are expressed in codes consisting of the three parameters Lightness-hue-Chroma, where Lightness is given in values between 1 and 5, the Chroma in values between 1 and 3, and the hue as one of “L”, “M”, or “R”.
  • a shade code in the Vita 3D-master can e.g. be 2M1, where the Lightness parameter equals 2, the Chroma 1 and the hue “M”.
  • the known color data of the VITA 3D-master shade guide reference shades can be provided by scanning the pre-manufactured teeth of the shade guide. These color data are then also initially obtained in the RGB color space and can be converted to the L*a*b color space using the same algorithms applied to the color information/data for the point P.
  • the known color data of each reference shade guide (having a code expressed in terms of Lightness, hue and Chroma) is then provided in terms of the L*a*b color space.
  • the value of the Lightness-parameter L* P in the point is determined first.
  • the value of L* P is compared with the values of the L*parameters for the reference tooth shades. If L* P is close to the L*-value for the i'th reference tooth shade value, L* Ri the L*parameter for point P may be set equal to L* Ri .
  • the Lightness parameter is not close to any of the references but instead is located almost in the middle between two L*-values.
  • the L*a*b color space is a linear space, the individual parameters of the shade values can be interpolated such that the Lightness for point P, L* P , can be set to 2.5.
  • a claim may refer to any of the preceding claims, and “any” is understood to mean “any one or more” of the preceding claims.
  • the features of the method described above and in the following may be implemented in software and carried out on a data processing system or other processing means caused by the execution of computer-executable instructions.
  • the instructions may be program code means loaded in a memory, such as a RAM, from a storage medium or from another computer via a computer network.
  • the described features may be implemented by hardwired circuitry instead of software or in combination with software.
  • Hassel 2012 Hassel et al. “Determination of VITA Classical shades with the 3D-Master shade guide”. Acta Ocdontol Scand. 2013; 71(3-4):721-6.
  • Dozic 2007 Dozic et al. “Performance of five commercially available tooth color-measuring devices”, J Prosthodont. 2007; 16(2):93-100.
  • a method for determining shade of a patient's tooth comprising:
  • determining the tooth shade value for the point comprises selecting the reference tooth shade value with the known texture value closest to the texture data of the point.
  • determining the tooth shade value for the point comprises an interpolation of the two or more reference tooth shade values having known texture values close to the texture data of the point.
  • the visual representation of the certainty score comprises a binary code, a bar structure with a color gradient, a numerical value, and/or a comparison between the texture data and the known texture value of the determined tooth shade value.
  • the one or more reference tooth shade values relate to shade values for natural teeth with intact surface and/or to shade values for teeth prepared for a dental restoration.
  • the method comprises comparing the texture data with known texture values for soft oral tissue, such as gum tissue and gingiva.
  • the texture information comprises at least one of tooth color or surface roughness.
  • the tooth shade profile comprises a one or more tooth shade regions on the tooth surface where an average tooth shade is derived for each region from tooth shade values determined for a number of points within the region.
  • obtaining the digital 3D representation of the tooth comprises recording a series of sub-scans of the tooth, where at least one of said sub-scans comprises both texture information and geometry information for said tooth, and generating the digital 3D representation of the tooth from the recorded series of sub-scans.
  • combining the texture information from the sub-scans comprises interpolating the texture information and/or calculating an average value of the texture information.
  • a user interface for determining and displaying shade of a patient's tooth, wherein the user interface is configured for:

Landscapes

  • Physics & Mathematics (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Epidemiology (AREA)
  • Dentistry (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Animal Behavior & Ethology (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Dental Tools And Instruments Or Auxiliary Dental Instruments (AREA)

Abstract

Disclosed in a method, a user interface and a system for use in determining shade of a patient's tooth, wherein a digital 3D representation including shape data and texture data for the tooth is obtained. A tooth shade value for at least one point on the tooth is determined based on the texture data of the corresponding point of the digital 3D representation and on known texture values of one or more reference tooth shade values.

Description

CROSS REFERENCE TO RELATED APPLICATIONS
The present application is a continuation of U.S. application Ser. No. 15/888,764, filed on Feb. 5, 2018, which is a continuation of U.S. application Ser. No. 15/117,078, filed on Aug. 5, 2016, now U.S. Pat. No. 10,010,387, which is a U.S. national stage of International Application No. PCT/EP2015/052537, filed on Feb. 6, 2015, which claims the benefit of Danish Application No. PA 2014-70066, filed on Feb. 7, 2014. The entire contents of each of U.S. application Ser. No. 15/888,764, U.S. application Ser. No. 15/117,078, International Application No. PCT/EP2015/052537, and Danish Application No. PA 2014-700665 are hereby incorporated herein by reference in their entirety.
TECHNICAL FIELD
This invention generally relates to methods and a user interfaces for determining the shade of a patient's tooth or teeth and for utilizing the determined tooth shades for designing and manufacturing dental restorations.
When designing and manufacturing a dental restoration for a patient, such as a crown or a bridge restoration, it is advantageous that both the shape and shade of the manufactured restoration is adapted to the patient's natural teeth surrounding the restoration. If the shade of the restoration differs significantly from the surrounding natural teeth, e.g. is significantly darker or brighter than these, the restoration appear artificial and deteriorate the aesthetic impression of the patient's smile.
The tooth color can be represented in many different color spaces, such as the L*C*h*color space representing color in terms of Lightness, Chroma and Hue, or in the L*a*b*color space as described e.g. by Hassel et al. (Hassel 2012) and Dozic et al. (Dozic 2007). The L*a*b*color space has the advantage that it is designed to approximate human vision with the L*component closely matches human perception of lightness.
In order to aid dental technicians in their manual work of manufacturing a restoration which appears natural, the tooth colors are often expressed in terms of reference tooth shade values of a tooth shade system (often referred to as a tooth shade guide). Each reference tooth shade value in a tooth shade guide represents a predetermined and known tooth color value and often correspond to the color of commercially available ceramics for the production of dental restorations. This is e.g. the case for the VITA 3D-Master or the VITA Classic shade guides provided by VITA Zahnfabrik, Germany.
In the VITA 3D-Master system, tooth shades are expressed in codes referring to the L*C*h* color space, where each code is constructed according to (Lightness, hue, Chroma). One example of a tooth shade value is 3R1.5 where “3” refers to the lightness, “R” to the hue and “1.5” to the Chroma of the tooth. This allows the dentist to describe the shade of the patient's tooth in terms that a dental technician immediately understands, such that the technician will know from which ceramics he should manufacture the restoration to provide that it has the correct shade.
When manually determining which reference tooth shade value best matches the color of a patient's tooth, the dentist holds different pre-manufactured teeth of the shade guide at the tooth for comparison. Often a picture is taken with the pre-manufactured structures arranged at the teeth. The technician who produces the prosthetic then uses the picture in evaluating which ceramic must be used for the different parts of the restoration based on the picture. This process is both time consuming and inaccurate.
SUMMARY
Disclosed is a method for determining shade of a patient's tooth, wherein the method comprises:
    • obtaining a digital 3D representation of the tooth, where the digital 3D representation comprises shape data and texture data for the tooth; and
    • determining a tooth shade value for at least one point on the tooth based on the texture data of the corresponding point of the digital 3D representation and on known texture values of one or more reference tooth shade values.
Disclosed is a user interface for determining and displaying shade of a patient's tooth, wherein the user interface is configured for:
    • obtaining a digital 3D representation of the tooth, said digital 3D representation comprising shape data and texture data for the tooth;
    • displaying at least the shape data of the digital 3D representation such that the shape of the tooth is visualized in the user interface;
    • determining a tooth shade value for at least one point on the tooth based on the texture data of the corresponding point of the digital 3D representation and on known texture values of one or more reference tooth shade values; and
    • displaying the determined tooth shade value.
The texture data of the digital 3D representation expresses the texture of the tooth. The texture data can be a texture profile expressing the variation in the texture over the tooth. The shape data of the digital 3D representation expresses the shape of the tooth.
In some embodiments, the texture information comprises at least one of tooth color or surface roughness.
When the texture information comprises tooth color information, the texture data expressing a texture profile of the tooth may be color data expressing a color profile of the tooth, and the tooth shade value for a point on the tooth may be derived by comparing the color data of the corresponding point of the digital 3D representation with known color values of one or more reference tooth shade values.
Determining both the tooth shade value from a digital 3D representation comprising both shape data expressing the shape of the tooth and texture data expressing a texture profile of the tooth provides the advantage that shade and geometry information are directly linked. This e.g. advantageous e.g. in CAD/CAM dentistry where dental restorations are designed using Computer Aided Design (CAD) tools and subsequently manufactured from the design using Computer Aided Design (CAM) tools. The material used for the manufacture of the dental restoration can then be selected based on the determined tooth shade value.
In many cases, the dental restoration is manufactured with a shade profile where the shade differs from the incisal edge towards cervical end of the restoration. The disclosed invention allows the operator to determine tooth shade values for several points on the tooth such that a shade profile can be determined for the dental restoration. Multi-shaded milling blocks exits which mimics standard tooth shade profiles. Having the shape data and the tooth shade values linked via the digital 3D representation provides that the correct portion of the multi-shaded milling block can be milled out. The remaining portion of the multi-shaded milling block forming the dental restoration will then have a shape and shade profile which closely resembles that of a natural tooth.
In some embodiments, obtaining the digital 3D representation of the tooth comprises recording a series of sub-scans of the tooth, where at least one of said sub-scans comprises both texture information and geometry information for said tooth, and generating the digital 3D representation of the tooth from the recorded series of sub-scans.
When a plurality of the sub-scans comprise texture information, the texture data for the digital 3D representation can be derived by combining the texture information of the several sub-scans.
The recorded sub-scans comprise at least data of the tooth for which the shade is determined, but potentially also of the neighboring teeth such that for example the shape and location of the neighboring teeth can be taken into account when designing a dental restoration for the tooth. Texture information and texture data for the neighboring teeth can also be used to determine the shade value for the tooth, e.g. by interpolation of the shades determined for the neighbor teeth.
In some embodiments, the method comprises creating a shade profile for the tooth from shade values determined for one or more of points on the tooth.
The shade profile of natural teeth often has a brighter shade at the incisal edge of the tooth and gradually changes into a darker shade towards the cervical end of the tooth, i.e. the end at the patient's gingiva.
When the tooth shade value is determined for one point only the tooth shade profile may be generated based on knowledge of the normal tooth shade profile for that particular type of tooth and patient. This knowledge may relate to how the shade profile normally changes over the tooth, the age and gender of the patients, etc.
Often the profile will be based on tooth shades determined in several points on the tooth to provide the most reliable tooth shade profile.
In some embodiments the user interface is configured for creating a shade profile for the tooth from tooth shade values determined for one or more points on the tooth.
In some embodiments, the tooth shade profile can be created by interpolation of tooth shade values determined for points distributed over the tooth surface with some distance between the points. The tooth shade value for some parts of the tooth surface are then not derived directly from sub-scan texture information relating to these parts but from the determined tooth shade values for other parts/points on the tooth surface. A tooth shade profile for the entire labial/buccal surface of the tooth can thus be created from a selection of points on the surface providing a fast and often sufficiently accurate procedure for creating the tooth shade profile. The interpolation of the tooth shade values can be realized by an interpolation in each of the coordinates of the color space used to describe the tooth color.
In some embodiments, the tooth shade profile comprises a one or more tooth shade regions on the tooth surface where an average tooth shade is derived for each region from tooth shade values determined for a number of points within the region.
The tooth shade region can be defined by a structure encircling a portion of the tooth surface in the digital 3D representation, where either the operator or a computer implemented algorithm decides where each geometric structure is located on the digital 3D representation. Different shapes (e.g. circles, squares, or rectangles) and sizes (e.g. corresponding to a few millimeters) of the geometric structure can be used. The number of points within the geometrical structure can be increased to provide a more accurate measure of the shade or reduced to provide a faster calculation.
The average tooth shade value for a region can e.g. be derived as a weighted average where the tooth shade value for points in the center of the structure is assigned a higher weight than tooth shade value of points closer to the boundary.
The tooth surface can also be divided into a coronal, a middle and a cervical region. Some natural teeth has a shade profile which can be expressed by such a division and many dentists and dental technicians are familiar with such a division.
In some embodiments, tooth shade values are determined for a plurality of teeth, i.e. on parts of the digital 3D representation corresponding to two or more teeth, and a tooth shade value and/or a tooth shade profile for each of these teeth is created from the determined tooth shade values.
In some embodiments, the texture data at least partly are derived by combining the texture information from corresponding parts of a number of the sub-scans.
The digital 3D representation can be generated through registration of sub-scans into a common coordinate system by matching overlapping sections of sub-scans, i.e. the sections of the sub-scans which relate to the same region of the tooth. When two or more sub-scans also comprise texture information relating to the same region of the tooth, deriving the texture data for this region in the digital 3D representation can comprise combining the corresponding texture information, i.e. the texture information in the sub-scans corresponding to the same sections of the tooth.
Deriving the texture data based on texture information from two or more sub-scan can provide a more accurate measurement of the texture data. The texture information of one sub-scan for a particular region of the tooth may be unreliable e.g. due to the angle between the surface in this region and the scanner when this particular sub-scan was recorded. The combination of texture information from several sub-scans can provide a more reliable color.
In some embodiments, combining the texture information from the sub-scans comprises interpolating the texture information, i.e. texture information from parts of the sub-scans corresponding to a point on the tooth are interpolated to determine the texture data for that point.
Such an interpolation can provide that the determined texture data is more accurate e.g. in cases where the texture information for a point on the tooth is not linearly varying over the sub-scans such that a simple averaging will not provide the best result.
In some embodiments, combining the texture information from the sub-scans comprises calculating an average value of the texture information, i.e. texture data for a point on the digital 3D representation are determined by averaging the texture information of the sub-scans corresponding to that point on the tooth.
In some embodiments, the calculated average value is a weighted average of the texture information.
This approach has the advantage that the derived texture data of the digital 3D representation are not as sensitive to errors in the texture information of a single sub-scan.
Such errors can be caused by several factors. One factor is the angle between the optical path of the probe light at the tooth surface and the tooth surface itself. When utilizing e.g. the focus scanning technique, the texture data for a point on the tooth is preferably derived from a number of sub-scans where at least some of the sub-scans are recorded at different orientations of the scanner relative to the teeth. The sections of the sub-scans relating to this point are hence acquired at different angles relative to the tooth surface in this point.
A portion of a sub-scan recorded from a surface perpendicular to the optical path of the probe light at the tooth may be dominated by specular reflected light which does not describe the texture of the tooth but rather the spectral distribution of the probe light. A portion of a sub-scan recorded from a tooth surface almost parallel to the optical path is often quite weak and hence often provide an erroneous detection of the texture at that point.
In some embodiments, the texture information from parts of a sub-scan relating to a tooth surface which is substantially perpendicular or parallel to the optical path are assigned a low weight in the weighted averaging of the texture information to determine the texture data for the point.
The orientation of the scanner relative to the tooth when a sub-scan is acquired can be determined from the shape of the sub-scan. Parts of the sub-scan relating to tooth surfaces which are substantially parallel or perpendicular to the optical path can thus immediately be detected in the sub-scan such that the texture information of the corresponding parts are assigned at low weight when determining the texture data for this point from a series of sub-scans.
A specular reflection from the tooth often has an intensity which is significantly higher than that of e.g. diffuse light from surfaces which have an oblique angle relative to the optical path. In some cases the specular reflection will saturate the pixels of the image sensor used for the recording of the sub-scans.
In some embodiments, the method comprises detecting saturated pixels in the recorded sub-scans and assigning a low weight to the texture information of the saturated pixels when combining the texture information from the sub-scans, i.e. when calculating the weighted average of the texture information.
Specular reflection from a tooth surface may also be detected from a comparison between the spectrum of the light received from the tooth and that of the probe light. If these spectra a very similar it indicates that the tooth has a perfectly white surface which is not natural. Such texture information may thus be assigned a low weight in a weighted average of texture information.
In some embodiments determining the tooth shade value for the point comprises selecting the reference tooth shade value with known texture value closest to the texture data of the point.
When the texture data comprises color data, selecting the tooth shade value of the point can comprise calculating the color difference between the determined color data in the point and the color data of the reference tooth shade values. This difference can e.g. be calculated as a Euclidian distance in the used color space. As an example, Dozic et al. (Dozic 2007) describes that the Euclidian distance ΔE between two points (L*1, a*1, b*1) and (L*2, a*2, b*2) in the L*a*b*color space is given by:
Δ E = ( L 1 * - L 2 * ) 2 + ( a 1 * - a 2 * ) 2 + ( b 1 * - b 2 * ) 2 2
Selecting the tooth shade value can then comprise determining for which of the reference tooth shades the color difference, i.e. the Euclidian distance, is the smallest.
In some embodiments determining the tooth shade value for the point comprises an interpolation of the two or more reference tooth shade values having known texture values close to the texture data of the point.
This interpolation provides that the tooth shade can be represented with a more detailed solution than what is provided by the tooth shade standard used to describe the tooth shade. For instance when using a Lightness-Hue-Chroma code a tooth shade value of 1.5M2.5 can be determined for the tooth by interpolation of Lightness values of 1 and 2, and Chroma values of 2 and 3.
The tooth shade value can be displayed in a user interface e.g. together with the digital 3D representation of the tooth. If the digital 3D representation also contains parts relating to other teeth the tooth shade value for the tooth is preferably displayed at the tooth, such as at the point of the for which the tooth shade value has been determined.
The tooth shade value can also be represented as a color mapped onto the digital 3D representation.
When a dental restoration is designed based on the determined tooth shade value this can provide a visualization of how the restoration will appear together with neighboring teeth also contained in the digital 3D representation obtained by scanning the teeth.
In some embodiments, the method comprises deriving a certainty score expressing the certainty of the determined tooth shade value.
Deriving a certainty score for the determined tooth shade value provides the advantage that a measure of how accurate the determined value is can be displayed to the operator, preferably when the patient is still at the clinic such that further scanning can be performed if this is required to provide a more precise tooth shade value.
In some embodiments, the method comprises generating a visual representation of the certainty score and displaying this visual representation in a user interface.
In some embodiments, the method comprises generating a certainty score profile at least for a portion of the tooth, where the certainty scope profile represents the certainty scores for tooth shade values determined for a number of points on the tooth, such as for the values in a tooth shade profile for the tooth. The certainty score profile can be mapped onto the digital 3D representation of the tooth and visualized in a user interface. When the tooth shade profile also is mapped onto the tooth digital 3D representation the operator may be allowed to toggle between having the tooth shade profile and having the certainty scope profiled visualized on the digital 3D representation.
In some embodiments the visual representation of the certainty score is displayed together with or is mapped onto the digital 3D representation of the tooth.
In some embodiments, the method comprises comparing the derived certainty score with a range of acceptable certainty score values. This is done to verify that the certainty score is acceptable, i.e. that the determined tooth shade value is sufficiently reliable.
One boundary of the range can be defined by a threshold value. When a high certainty scope indicates that the determined shade value most likely is correct, the threshold value may define the lower boundary of the range and vice versa.
A visual representation of the certainty score or of the result of the comparison of the certainty score with the range can be generated and displayed in a user interface. Preferably, this visual representation is displayed together with the determined tooth shade value.
In some embodiments, the method comprises deciding based on the certainty score whether the determined tooth shade value or tooth shade profile is acceptable. This may be based on the comparison of the derived certainty score and the range of acceptable certainty score values, e.g. where it is decided that the determined tooth shade value is acceptable if the certainty score is within the range of acceptable values.
In some embodiments, the certainty measure relates to how uniform the sub-scan texture information is at the point.
If large variations are found in the texture information in the vicinity of the parts corresponding to the point for a substantial fraction of the sub-scans, the texture data derived therefrom may be unreliable and the tooth shade value derived for this point is accordingly not very reliable.
In some embodiments, the certainty measure relates to how close the texture data is to the known texture value of the determined tooth shade value. In particular, the certainty measure may relate to how close one parameter of the color data of the digital 3D representation is to the corresponding parameter of the known color for the determined tooth shade value. For example, the certainty measure may relate to the difference in the lightness parameter between point of the digital 3D representation and the determined tooth shade value.
The Euclidian distance between the color data to the selected reference tooth shade value can also be used in determining the certainty measure. If the Euclidian distance is above a threshold value the uncertainty is then evaluated to be too large. The color data can here both relate to color data of the point or the average color data for a region surrounding the point.
In some embodiments, the certainty measure relates to the amount of texture information used to derive the texture data at the point.
When the texture data for the point is derived from a limited amount of texture information the texture data, and accordingly the tooth shade value derived therefrom, may be less reliable than the tooth shade values derived from large amounts of texture information.
In some embodiments, the visual representation of the certainty score comprises a binary code, such as red for certainty scores outside a range of acceptable certainty score values, and green for certainty scores within the range, a bar structure with a color gradient, a numerical value, and/or a comparison between the texture data and the known texture value of the determined tooth shade value.
In some embodiments, the visual representation of the certainty score comprises a certainty score indicator.
The certainty score indicator may comprise a bar structure with a color gradient going from a first color representing a low certainty score to a second color representing a high certainty score. The first color may be red and the second color green. The color gradient of the bar structure may be configured to have an intermediate color, e.g. yellow representing the threshold value for the certainty score. The certainty score indicator may comprise marker which is arranged relative to the color gradient of the bar structure such that it indicated the certainty score.
In some embodiments, the visual representation of the certainty score comprises a numerical value, such as a numerical value in an interval extending from a lower limit indicating a low certainly, i.e. a relatively uncertain tooth shade value, to a higher limit indicating a high certainty, i.e. a relatively certain tooth shade value.
In some embodiments, the one or more reference tooth shade values relate to shade values for natural teeth with intact surface and/or to shade values for teeth prepared for a dental restoration.
The reference tooth shade values used for determining the tooth shade can be selected based on the tooth. Intact and healthy teeth normally have tooth shades in one range of tooth shade values where a tooth prepared for a dental restoration has a tooth shade in another range, which may overlap with the range for healthy teeth. It may thus be advantageous that the operator enters whether the tooth is intact or prepared for a restoration and the appropriate color space is used in the comparison with the texture data.
If the color data in the point on the digital 3D representation of the tooth has a poor match to all the reference tooth shade values of the selected tooth shade system/guide the point may e.g. be on the gingiva of the patient or relate to silver filling.
In some embodiments, the method comprises comparing the texture data with known texture values for soft oral tissue, such as gum tissue and gingiva.
This may e.g. be relevant when the certainty scores are outside said range of acceptable certainty score values for all tooth shade values of a tooth shade system, i.e. if there is a poor match between the texture data and the known texture for all the tooth shades of the reference set.
In a user interface for implementing the method, it may be suggested to the operator that the point perhaps is not on a tooth surface but on the gums or gingiva of the patient. This suggestion may be provided both when the texture data has been found to give a good match with known texture values of gum/gingiva and/or when the texture data has a poor match with the known texture values of the reference tooth shade values in the tooth shade system or systems.
In some embodiments, the method comprises determining an alternative tooth shade value for the point when said certainty score is outside said range of acceptable certainty score values.
In some embodiments, the method comprises displaying the alternative tooth shade value in the user interface optionally together with the digital 3D representation of the patient's set of teeth and/or the initially determined tooth shade value.
The digital 3D representation of the tooth is generated at least partly from the geometry information of the sub-scans. In some embodiments, the texture information of the sub-scans is also taken into account when generating the digital 3D representation of the tooth.
Sub-scans comprising texture information and geometry information may be recorded for more than said tooth, such that the generated digital 3D representation may comprise shade data expressing the shape and texture data expressing the texture profile of several of the patient's teeth.
Disclosed is a method for determining shade of a patient's tooth, wherein the method comprises:
    • recording a series of sub-scans of the patient's set of teeth, where a plurality of said sub-scans comprises both texture information and geometry information for said tooth;
    • generating a digital 3D representation of the tooth from said sub-scans, wherein the digital 3D representation comprises shape data expressing the shape of the tooth and texture data expressing a texture profile of the tooth; and
    • determining a tooth shade value for a point on the tooth by comparing the texture data of the corresponding point of the digital 3D representation with a known texture value of one or more reference tooth shade values.
Disclosed is a user interface for determining and displaying shade of a patient's tooth, wherein the user interface is configured for:
    • obtaining a digital 3D representation of the tooth, said digital 3D representation comprising shape data expressing the shape of the tooth and texture data expressing a texture profile of the tooth;
    • displaying at least the shape data of the digital 3D representation such that the shape of the tooth is visualized in the user interface;
    • determining a tooth shade value for a point on the tooth by comparing the texture data of the corresponding point of the digital 3D representation with a known texture value of one or more reference tooth shade values; and
    • displaying the determined tooth shade value.
In some embodiments, the user interface is configured for deriving a certainty score expressing the certainty of the determined tooth shade value for said point.
In some embodiments, the user interface comprises a virtual tool which when activated on a point of the digital 3D representation of the tooth provides that
    • the determined tooth shade value for the point; and/or
    • a visual representation of a certainty score for the determined tooth shade value; and/or
    • a visual representation of a comparison of the derived certainty score with a range of acceptable certainty score values
      is visualized in the user interface.
The user interface can then provide the operator with an opportunity to decide based on the visualized certainty score and/or the visual representations whether the determined tooth shade value or tooth shade profile is acceptable.
In some embodiments, the visual representation of the comparison of the derived certainty score with the range of acceptable certainty score values comprises a binary code, such as red for certainty scores outside a range of acceptable certainty score values, and green for certainty scores within the range. Other means for this visualization are described above.
The visualized certainty score and/or the representation(s) of the certainty score or comparison of the certainty score with the range of acceptable certainty score values may be displayed at the digital 3D representation in the user interface or in a shade value region of the user interface.
In some embodiments, the user interface is configured for determining an alternative shade value for the point and for displaying the alternative shade value when the certainty scores outside a range of acceptable certainty score values.
Disclosed is a method for designing a dental restoration for a patient, wherein the method comprises:
    • obtaining a digital 3D representation of at least one tooth, said digital 3D representation comprising shape data expressing the shape of the tooth and texture data expressing a texture profile of the tooth;
    • determining a tooth shade value for a point on the tooth by comparing the texture data of the corresponding point of the digital 3D representation with a known texture value of one or more reference tooth shade values; and
    • creating a digital restoration design for one or more of the patient's teeth; and
    • selecting a restoration shade of the digital restoration design based on said tooth shade value.
The digital restoration design can e.g. be for the manufacture of dental prosthetic restoration for the patient, such as a crown or a bridge restoration, where the digital restoration design expresses a desired shape and shade profile of the dental restoration. Such digital restoration designs can be in the form of a CAD model of the dental restoration.
In some embodiments, the method comprises suggesting a dental material for manufacturing the dental restoration from the digital restoration design based on the determined restoration shade.
In cases where the dental restoration is designed and manufactured for an existing tooth which has an acceptable shade, the tooth shade value or tooth shade profile can be determined for the existing tooth and the shade of the digital restoration design based on the tooth shade value or tooth shade profile of the existing tooth.
This may e.g. be advantageous for the crown portions of a bridge restoration in the case where the tooth which is intended to accept the crown portion of the bridge is a healthy tooth.
In some cases the dental restoration is designed and manufactured for a tooth which either is damaged or has an undesired shade profile, such as for a broken or dead tooth. In such cases it can be advantageous to determine the tooth shade value or tooth shade profile for one or more of the neighboring teeth and selecting the restoration shade of the digital restoration design from e.g. an interpolation of the tooth shade values/profiles of the neighboring teeth.
Disclosed is a method for designing a dental restoration for a first tooth, wherein the method comprises:
    • obtaining a digital 3D representation of the patient's set of teeth, said digital 3D representation comprising shape data and texture data expressing the shape and texture profile, respectively, of at least one second tooth;
    • designing a digital restoration design for the first tooth;
    • deriving a desired texture profile of the digital restoration design from the texture data of the at least one second tooth; and
    • determining a restoration shade value or restoration shade profile of the digital restoration design by comparing the desired texture profile with texture values for one or more reference tooth shade values.
In some embodiments, the desired texture profile is derived by interpolation or averaging of the texture data of the digital 3D representation of the neighbor teeth.
In some embodiments, one or more of the sub-scans comprise texture information for the patient's soft tissue, and optionally geometry information for said soft tissue. The generated digital 3D representation may then comprise shape data expressing the shape of the soft tissue and texture data expressing a texture profile of the soft tissue.
From this information, an aesthetical pleasing denture can be designed where the color of the soft tissue part of the denture is selected based on the texture profile of the corresponding part of the digital 3D representation.
Knowledge of the texture of the soft tissue, such as of the color of the soft tissue, can also be used for diagnostics. When the texture data of a point on the digital 3D representation corresponding to soft tissue does not provide a sufficiently good match with a known range of texture values for soft tissue, a warning may be prompted in a user interface to alert the operator that the soft tissue is suspicious.
Disclosed is a system for determining shade of a patient's tooth, wherein the system comprises:
    • a scanner capable of recording a digital 3D representation of the tooth, where the digital 3D representation comprises shape data and texture data for the tooth; and
    • a data processing system comprising a computer-readable medium having stored thereon the program code means for causing the data processing system to determine a tooth shade value for at least one point on the tooth based on the texture data of the corresponding point of the digital 3D representation and on known texture values of one or more reference tooth shade values using the method according to any of the embodiments.
In some embodiments, the sub-scans are recorded using an intra-oral scanner, such as the 3Shape TRIOS intra-oral scanner.
The intra-oral scanner may be configured for utilizing focus scanning, where the sub-scans of the scanned teeth are reconstructed from in-focus images acquired at different focus depths. The focus scanning technique can be performed by generating a probe light and transmitting this probe light towards the set of teeth such that at least a part of the set of teeth is illuminated. Light returning from the set of teeth is transmitted towards a camera and imaged onto an image sensor in the camera by means of an optical system, where the image sensor/camera comprises an array of sensor elements. The position of the focus plane on/relative to the set of teeth is varied by means of focusing optics while images are obtained from/by means of said array of sensor elements. Based on the images, the in-focus position(s) of each of a plurality of the sensor elements or each of a plurality of groups of the sensor elements may be determined for a sequence of focus plane positions.
The in-focus position can e.g. be calculated by determining the maximum of a correlation measure for each of a plurality of the sensor elements or each of a plurality of groups of the sensor elements for a range of focus planes as described in WO2010145669. From the in-focus positions, sub-scans of the set of teeth can be derived with geometry information relating to the shape of the scanned surface. When e.g. the image sensor is a color sensor and the light source provides a multispectral signal a plurality of the sub-scans can include both geometry information and texture information, such as color information, for said tooth.
A digital 3D representation of the set of teeth can then be generated from the recorded sub-scans by e.g. the use of an Iterative Closest Point (ICP) algorithm. Iterative Closest Point (ICP) is an algorithm employed to minimize the difference between two clouds of points. ICP can be used to reconstruct 2D or 3D surfaces from different scans or sub-scans. The algorithm is conceptually simple and is commonly used in real-time. It iteratively revises the transformation, i.e. translation and rotation, needed to minimize the distance between the points of two raw scans or sub-scans. The inputs are: points from two raw scans or sub-scans, initial estimation of the transformation, criteria for stopping the iteration. The output is: refined transformation. Essentially the algorithm steps are:
    • 1. Associate points by the nearest neighbor criteria. 2. Estimate transformation parameters using a mean square cost function. 3. Transform the points using the estimated parameters. 4. Iterate, i.e. re-associate the points and so on.
The generated digital 3D representation formed by such a procedure comprises shape data expressing the shape of the tooth. The texture information of the sub-scans can be used in various ways to provide that the generated digital 3D representation also comprises texture data expressing a texture profile of the tooth. For a number of the sub-scans, the part of the sub-scan relating to the same point on the tooth can be identified, e.g. during the ICP procedure. The corresponding texture information of these parts of the sub-scans can then be combined to provide the texture data for that point.
Furthermore, the invention relates to a computer program product comprising program code means for causing a data processing system to perform the method according to any of the embodiments, when said program code means are executed on the data processing system, and a computer program product, comprising a computer-readable medium having stored there on the program code means.
The present invention relates to different aspects including the method and user interface described above and in the following, and corresponding methods and user interface, each yielding one or more of the described advantage, and each having one or more embodiments corresponding to the embodiments described above and/or disclosed in the appended claims.
BRIEF DESCRIPTION OF THE DRAWINGS
The above and/or additional objects, features and advantages of the present invention, will be further elucidated by the following illustrative and non-limiting detailed description of embodiments of the present invention, with reference to the appended drawings, wherein:
FIG. 1 shows an example of a flow chart for an embodiment.
FIGS. 2 to 4 show parts of screen shots of user interfaces.
FIG. 5 shows steps of a method for designing a dental restoration.
FIG. 6 shows a schematic of a system for determining tooth shade values.
FIGS. 7A-7D and 8A-8B show schematics of intra-oral scanning.
FIGS. 9A-9B illustrates one way of determining tooth shade values from texture data.
DETAILED DESCRIPTION
In the following description, reference is made to the accompanying figures, which show by way of illustration how the invention may be practiced.
FIG. 1 shows an example of a flow chart 100 for an embodiment of the method for determining shade of a patient's tooth.
In step 102 a series of sub-scans of the patient's set of teeth is recorded, where a plurality of said sub-scans comprises both texture information and shape information for the tooth.
In step 103 a digital 3D representation of the tooth is generated from said sub-scans, where the digital 3D representation comprises texture data expressing a texture profile of the tooth. The digital 3D representation further comprises shape data expressing the shape of the tooth such that the shape of the tooth can be visualized in a user interface.
In step 104 a tooth shade value for a point on the tooth is determined based on the texture data. This is done at least in part by comparing the texture data of the corresponding point of the digital 3D representation with a known texture value of one or more reference tooth shade values. The reference tooth shade values may be provided in the form of a library file and comprise tooth shade values and corresponding texture values based on e.g. the VITA 3D-Master and/or the VITA Classic tooth shade systems.
FIGS. 2 to 4 show parts of screen shots from user interfaces in which derived tooth shade values and visual representation of the corresponding certainty scores for a number of tooth regions are displayed at the digital 3D representations of the patient's set of teeth.
The point or points on the tooth for which the tooth shade value(s) is/are determined can be selected by an operator. This can be the case e.g. when the digital 3D representation of the tooth is visualized in a user interface and the operator uses a pointing tool, such as a computer mouse, to indicate where on the digital 3D representation of the tooth, he wishes to determine the tooth shade value. The point or points can also be selected by a computer implemented algorithm based on predetermined positions on the digital 3D representation of the tooth, such as a point arranged at a certain distance to the incisal edge of the tooth.
The screen shot 210 seen in FIG. 2 shows three regions 212, 213, 214 on the digital 3D representation of the patient's set of teeth. Two of these 212, 213, are selected at the part of the digital 3D representation corresponding to the tooth 211 while the third 214 is selected on the soft tissue part 215 of the digital 3D representation. Average tooth shade value for a region can be calculated by averaging over tooth shade values derived for a number of points within the region or by calculating an average texture value for the region and determining the average tooth shade value therefrom. The average tooth shade values are displayed in tooth value sections 217, 218, 219 linked to the regions in the user interface. In the tooth value sections 217, 218 relating to the regions 212, 213 two tooth shade values are displayed where the upper shade value is derived using known texture values corresponding to the reference tooth shade values of the VITA 3D-Master tooth shade system and the lower tooth shade values relates to the VITA Classic tooth shade system. It is also seen that for the region 213 closest to the gingiva, the tooth shade is determined to be 2L1.5 in the VITA 3D-Master system and B1 in the VITA Classic system. In FIG. 2 the certainty scores for the derived tooth shade values are visualized as a certainty score indicator displayed next to the tooth shade values. In FIG. 2 the visualization of the certainty score indicator is in the form of a checkmark which indicates that the certainty score is sufficiently good to provide that the derived tooth shade values can be relied upon. The color of the checkmark may provide further information to the certainty score, such as in cases where a green checkmark indicates a more certain tooth shade value than a yellow checkmark. The third region 214 is located at the patient's soft tissue. An anatomical correct tooth shade value can hence not be calculated from the texture data of that part of the digital 3D representation of the patient's teeth and the corresponding certainty scope is accordingly very low. The visualization of the certainty score in the tooth value section 219 is hence a cross indicating that the derived shade value was rejected. Further no shade value is indicated in the tooth value section 219.
The screen shot 310 seen in FIG. 3 shows two regions 312, 314 on the digital 3D representation of the patient's set of teeth. One of these regions 312 is selected at the part of the digital 3D representation corresponding to the tooth 311 while the second region 314 is selected on the soft tissue part 315 of the digital 3D representation. Average tooth shade value for a region can be calculated as described above in relation to FIG. 2 . Shade value sections 317, 319 are also displayed for the regions 312, 314. Two tooth shade values 321 are derived for the region 312 and displayed in the corresponding tooth value section 317, where the upper value is derived using known texture values corresponding to the reference tooth shade values of the VITA 3D-Master tooth shade system (derived tooth shade value is 1.5M1) and the lower value using the VITA Classic tooth shade system (derived tooth shade value is B1). In FIG. 3 the certainty score is visualized in the form of a certainty score indicator 322 comprising a vertical bar with a color gradient going from red representing a poor certainty score to green representing a good certainty score. The certainty score indicator has a marker indicating the certainty score on the bar. It is seen that the tooth shade value 1.5M1 of the VITA 3D-Master system is more certain than the tooth shade value B1 of the VITA Classic system for this region. The tooth shade value of 1.5M1 is found by interpolation of the reference tooth shades 1M1 and 2M2.
The second region 314 is located at the patient's soft tissue. An anatomical correct tooth shade value can hence not be calculated from the texture data of that part of the digital 3D representation of the patient's teeth and the corresponding certainty scope is accordingly very low as seen in the vertical bars of tooth value section 319.
FIG. 4 shows a screen shot 410 where determined tooth shade values are derived for a total of 15 regions on the digital 3D representation of the tooth 411. The tooth shade values are all derived based on the known texture values of the reference tooth shade values of the VITA 3D-Master tooth shade system. The certainty scores are visualized in the form of a certainty score indicator comprising a vertical bar with a color gradient going from red representing a poor certainty score to green representing a good certainty score. As can be seen in the tooth value sections 417, 418 of the user interface there are large variations in the certainty scores. For example, the certainty score for the region 412 is almost at maximum while the certainty score of the region 413 is much close to a threshold for acceptable certainty score values. When tooth shade values are determined for a number of points on the tooth, the points may be arranged in a grid over the part of the digital 3D representation of the tooth.
FIG. 5 shows steps of a method for designing a dental restoration.
In step 531 a digital restoration design is created e.g. based on the shape data of a digital 3D representation of the patient's set of teeth and/or on template digital restoration design loaded from a library. Template digital restoration designs may e.g. be used when the tooth is broken.
In step 532 the tooth shade values of different points or regions of the teeth are derived from the texture data of the digital 3D representation of the patient's set of teeth. From the derived tooth shade values or from tooth shade profiles created based on the derived tooth shade values a desired shade profile for the dental restoration can be determined. This can be based on e.g. feature extraction where shade values are extracted from the other teeth by e.g. identifying shade zones on these teeth and copying these zones to the dental restoration. It can also be based on established shade rules for teeth, e.g. a rule describing a relation between the tooth shades values or profiles of the canines and the anterior teeth.
In step 533 the desired tooth shade value(s) for the dental restoration is merged into the digital restoration design.
When the dental restoration is to be drilled from a multicolored milling block it is important that the dental restoration is milled from the correct parts of the milling block. In step 534 a CAD model of the milling block is provided, where the CAD model comprises information of the shade profile of the milling block material. The optimal position of the digital restoration design relative to the CAD model of the milling block is then determined in 535, where different criteria can be apply to provide the best fit between the desired shade profile and what actually can be obtained as dictated by the shade profile of the milling block.
In step 536 the dental restoration is manufactured from the milling block by removing milling block material until the dental restoration is shaped according to the digital restoration design.
FIG. 6 shows a schematic of a system for determining tooth shade values. The system 640 comprises a computer device 642 comprising a computer readable medium 643 and a processor 644. The system further comprises a visual display unit 647, a computer keyboard 645 and a computer mouse 646 for entering data and activating virtual buttons in a user interface visualized on the visual display unit 647. The visual display unit can be a computer screen. The computer device 642 is capable of receiving a digital 3D representation of the patient's set of teeth from a scanning device 641, such as the TRIOS intra-oral color scanner manufactured by 3 shape A/S, or capable of receiving scan data from such a scanning device and forming a digital 3D representation of the patient's set of teeth based on such scan data. The obtained digital 3D representation can be stored in the computer readable medium 643 and provided to the processor 644. The processor is configured for implementing the method according to any of the embodiments. This may involve presenting one or more options to the operator, such as where to derive the tooth shade value and whether to accept a derived tooth shade value. The options can be presented in the user interface visualized on the visual display unit 647.
Many scanning devices have Bayer color filters with Red, Green and Blue filters and hence record color information in the RGB color space. For instance a focus scanner can record series of 2D color images for the generation of sub-scans, where the color information is provided in the RGB color space. The processor 644 then comprises algorithms for transforming the recorded color data into e.g. the L*a*b or L*C*h color spaces.
The system may further comprise a unit 648 for transmitting a digital restoration design and a CAD model of a milling block to e.g. a computer aided manufacturing (CAM) device 649 for manufacturing a shaded dental restoration or to another computer system e.g. located at a milling center where the dental restoration is manufactured. The unit for transmitting the digital restoration design can be a wired or a wireless connection.
The scanning of the patient's set of teeth using the scanning device 641 can be performed at a dentist while deriving the tooth shade values can be performed at a dental laboratory. In such cases the digital 3D representation of the patient's set of teeth can be provided via an internet connection between the dentist and the dental laboratory.
FIGS. 7A-7D and 8A-8B show schematics of intra-oral scanning.
Different scanner configurations can be used to acquire sub-scans comprising both shape and texture information. In some scanner designs the scanner is mounted on axes with encoders which provides that the sub-scans acquired from different orientations can be combined using position and orientation readings from the encoders. When the scanner operates by the focus-scanning technique the individual sub-scans of the tooth are derived from a sequence of 2D images obtained while scanning a focus plane over a portion of the tooth. The focus scanning technique is described in detail in WO2010145669. The shape information of the sub-scans for an object, such as a tooth, can be combined by algorithms for stitching and registration as widely known in the literature. Texture data relating to the tooth color can be obtained using a scanner having a multi-chromatic light source, e.g. a white light source and a color image sensor. Color information from multiple sub-scans can be interpolated and averaged by methods such as texture weaving, or by simply averaging corresponding color components of the sub-scans corresponding to the same point/location on the surface. Texture weaving is described by e.g. Callieri M, Cignoni P, Scopigno R. “Reconstructing textured meshes from multiple range rgb maps”. VMV 2002, Erlangen, Nov. 20-22, 2002.
In FIG. 7A the scanner 741 (here represented by a cross-sectional view of the scanner tip) is held in one position relative to the teeth 711, 760 (also represented by a cross-sectional view) while recording a sequence of 2D images for one sub-scan. The illustrated teeth can e.g. be the anterior teeth in the lower jaw. The size of the Field of View (here represented by the full line 761 on the teeth) of the scanner is determined by the light source, the optical components and the image sensor of the scanner. In the illustrated example, the Field of View 761 covers part of the surface of the tooth 711 and part of the surface of the neighbor tooth 760. The generated digital 3D representation can thus also contain data for the neighbor teeth. This is often advantageous, e.g. when the generated digital 3D representation is used for creating a digital restoration design for the manufacture of a dental restoration for the tooth. In the Figure, the scanner is arranged such that the acquired sub-scan comprises shape and color information for the incisal edge 762 of the teeth. The probe light rays 763 from the scanner corresponding to the perimeter of the Field of View are also shown in the Figure. These probe light rays 763 define the optical path 764 of the scanner probe light at the tooth 711.
A digital 3D representation of the tooth can be generated by combining sub-scans acquired from different orientations relative to the teeth, e.g. by sub-scan registration. Sub-scans acquired from three such different orientations are illustrated in FIGS. 7B, 7C and 7D, where only the optical path 763 of the scanner probe light is used to represent the relative scanner/tooth orientation in FIGS. 7C and 7D. The sub-scans (here represented by the full line 765 on the teeth) covers different but overlapping sections of the tooth surface such that the sub-scans can be combined by registration into a common coordinate system using e.g. an Iterative Closest Point (ICP) algorithm as described above. A segment of each of the sub-scans corresponds to the point P on the tooth surface. When the sub-scans are registered to generate a digital 3D representation of the tooth, a correlation between these segments is established and the texture information of these sub-scan segments can be combined to determine the texture data for point P on the generated digital 3D representation of the tooth.
One way of doing this is to calculate the average value for each of the parameters used to describe the texture. For example, when the L*a*b*color system is used to describe the color information provided in each sub-scan, the color data of the digital 3D representation can be derived by averaging over each of the L*, a*, and b*parameters of the sub-scans. For example, the L*parameter of the color data for a given point P is then given by
L * ( P ) = 1 N i N L i * ( P )
where N is the number of sub-scans used in deriving the texture data and L*i(P) is the L*parameter of the i'th sub-scan for the segment relating to P. Equivalent expressions are true for the a*and b*parameters for point P. The color parameters for each point on the digital 3D representation of the tooth can be determined for sections of or the entire surface of the tooth, such that the generated digital 3D representation comprises both shape and texture information about the tooth. The spatial resolution of the color data does not necessarily have to be identical to the resolution of the shape data of the digital 3D representation. The point P can be described e.g. in Cartesian, cylindrical or polar coordinates.
When the color data is derived for a point on the tooth, the tooth shade value for that point can be determined by comparing the derived color data with the known color data of the reference tooth shade values of a tooth shade guide such as the VITA 3D-Master.
FIG. 8A-8B illustrates some potentially problematic tooth surface areas for particular arrangements of the scanner 841 relative to the tooth 811.
FIG. 8A shows two points Pi and Pii on the tooth 811 where the tooth surface is either substantially perpendicular or parallel to the optical path, such that the texture information recorded at Pi and Pii may be unreliable. This is because the tooth surface at Pi is perpendicular to the optical path 864 i at point Pi which introduces the risk of having specular reflections of the probe light. The optical path 864 ii at point Pii is parallel to the tooth surface at Pii such that the signal recorded from this part of the tooth surface in this sub-scan is relatively weak. This may cause that the color information in this section of the sub-scan are unreliable.
In order to obtain more precise color data the averaging of the color information described above in relation to FIG. 7 can be a weighted averaging where the color information of unreliable sub-scans segments are assigned a lower weight than others.
In FIG. 8B is indicated three different optical paths 864 i, 864 ii and 864 iii at which sub-scans are acquired. When combining the color information for point P the color information of the segments of the sub-scans recorded with optical paths 864 i and 864 ii should be given a lower weight that the color information of the segment of the sub-scan recorded with the optical path 864 iii.
This can be expressed by a modification of the equation given above. For a weighted averaging of the color information, the L*parameter of the color data for a given point P is given by L*(P)=Σi Ni(P)·L*i(P)}/Σi Nαi where αi(P) is the weight factor for the color information of the i'th sub-scan in the segment at P. When a given sub-scan (e.g. the j'th sub-scan) is recorded at an angle relative to the tooth surface which causes the optical path to be e.g. perpendicular to the tooth surface at P, the corresponding weight factor αi(P) is given a lower value than the color data of sub-scans acquired with an oblique angle between the optical path and the tooth surface.
Equivalent equations are true for the a*and b*parameters of the color data for point P.
FIG. 9A-9B illustrates how a tooth shade value for a point P on a tooth can be determined based on reference tooth shade values.
For a given point P on the digital 3D representation, the color data (L*P, a*P, b*P) has been determined, e.g. by combining the color information of a series of sub-scans used for generating the digital 3D representation. If the color information originally is recorded using the RGB color space it is transformed into the L*a*b*color space using algorithms known to the skilled person.
In the example illustrated by FIG. 9A, the color data of the digital 3D representation and the known color values of the reference tooth shades are expressed in the L*a*b*color space, and the reference tooth shades are those from the VITA classical shade guide.
The reference shade values of the Vita classical shade guide are: B1, A1, B2, D2, A2, C1, C2, D4, A3, D3, B3, A3.5, B4, C3, A4, and C4. The color data of these reference shades can be provided by scanning the corresponding pre-manufactured teeth of the shade guide. These color data are then also initially obtained in the RGB color space and can be converted to the L*a*b color space using the same algorithms applied to the color information/data for the point P.
The tooth shade value for the point is determined as the reference tooth shade value which has the smallest Euclidian distance to the point in the L*a*b color space. The Euclidian distance ΔEP-R 1 from the color (L*P, a*P, b*P) to the known colors of the reference tooth shade values are calculated using the expression:
Δ E P - R i = ( L P * - L R i * ) 2 + ( a P * - a R i * ) 2 + ( b P * - b R i * ) 2 2
where Ri refers to the i'th reference tooth shade.
In FIG. 9A only the known colors (L*R1, a*R1, b*R1) and (L*R2, a*R2, b*R2) for the two closest reference values R1 and R2, respectively, are illustrated for simplicity. It can be seen that the Euclidian distance in the color space from P to R2 is the smallest, and the tooth shade in point P is hence selected as that of R2.
The certainty score for the tooth shade value determined for point P depends on how close the color data of the point P is to the known color value of the selected reference tooth shade value. This can be quantified by the Euclidian distance and since point P is not particularly close to R2 in FIG. 9A the determined tooth shade has a poor certainty value.
An alternative approach to using the Euclidian distance is to determine individual parameters of the tooth shade value one at a time. This approach can be used e.g. when the reference tooth shades values are those of the Vita 3D-master system.
The reference tooth shade values of the Vita 3D-master shade guide are expressed in codes consisting of the three parameters Lightness-hue-Chroma, where Lightness is given in values between 1 and 5, the Chroma in values between 1 and 3, and the hue as one of “L”, “M”, or “R”. A shade code in the Vita 3D-master can e.g. be 2M1, where the Lightness parameter equals 2, the Chroma 1 and the hue “M”.
The known color data of the VITA 3D-master shade guide reference shades can be provided by scanning the pre-manufactured teeth of the shade guide. These color data are then also initially obtained in the RGB color space and can be converted to the L*a*b color space using the same algorithms applied to the color information/data for the point P. The known color data of each reference shade guide (having a code expressed in terms of Lightness, hue and Chroma) is then provided in terms of the L*a*b color space.
Since the lightness L has the largest impact on the human perception of the tooth color, the value of the Lightness-parameter L*P in the point is determined first. The value of L*P is compared with the values of the L*parameters for the reference tooth shades. If L*P is close to the L*-value for the i'th reference tooth shade value, L*Ri the L*parameter for point P may be set equal to L*Ri.
In some cases the Lightness parameter is not close to any of the references but instead is located almost in the middle between two L*-values. For example when L*P in the point is between the values of L*Ri=2 and L*Ri+1=3 with almost equal distance to each of these as illustrated in FIG. 9B. Since the L*a*b color space is a linear space, the individual parameters of the shade values can be interpolated such that the Lightness for point P, L*P, can be set to 2.5.
The same procedure is performed for first the Chroma parameter and finally for the hue such that the three parameter of the tooth shade value are determined.
Although some embodiments have been described and shown in detail, the invention is not restricted to them, but may also be embodied in other ways within the scope of the subject matter defined in the following claims. In particular, it is to be understood that other embodiments may be utilized and structural and functional modifications may be made without departing from the scope of the present invention.
In device claims enumerating several means, several of these means can be embodied by one and the same item of hardware. The mere fact that certain measures are recited in mutually different dependent claims or described in different embodiments does not indicate that a combination of these measures cannot be used to advantage.
A claim may refer to any of the preceding claims, and “any” is understood to mean “any one or more” of the preceding claims.
It should be emphasized that the term “comprises/comprising” when used in this specification is taken to specify the presence of stated features, integers, steps or components but does not preclude the presence or addition of one or more other features, integers, steps, components or groups thereof.
The features of the method described above and in the following may be implemented in software and carried out on a data processing system or other processing means caused by the execution of computer-executable instructions. The instructions may be program code means loaded in a memory, such as a RAM, from a storage medium or from another computer via a computer network. Alternatively, the described features may be implemented by hardwired circuitry instead of software or in combination with software.
REFERENCES
Hassel 2012: Hassel et al. “Determination of VITA Classical shades with the 3D-Master shade guide”. Acta Ocdontol Scand. 2013; 71(3-4):721-6.
Dozic 2007: Dozic et al. “Performance of five commercially available tooth color-measuring devices”, J Prosthodont. 2007; 16(2):93-100.
EMBODIMENTS
1. A method for determining shade of a patient's tooth, wherein the method comprises:
    • obtaining a digital 3D representation of the tooth, where the digital 3D representation comprises shape data and texture data for the tooth; and
    • determining a tooth shade value for at least one point on the tooth based on the texture data of the corresponding point of the digital 3D representation and on known texture values of one or more reference tooth shade values.
2. The method according to embodiment 1, wherein determining the tooth shade value for the point comprises selecting the reference tooth shade value with the known texture value closest to the texture data of the point.
3. The method according to embodiment 1 or 2, wherein determining the tooth shade value for the point comprises an interpolation of the two or more reference tooth shade values having known texture values close to the texture data of the point.
4. The method according to any one of the preceding embodiments, wherein the method comprises deriving a certainty score expressing the certainty of the determined tooth shade value.
5. The method according to embodiment 4, wherein the method comprises generating a visual representation of the certainty score and displaying this visual representation in a user interface.
6. The method according to embodiment 5, wherein the visual representation of the certainty score is displayed together with or is mapped onto the digital 3D representation of the tooth.
7. The method according to any one of embodiments 4 to 6, wherein the method comprises comparing the derived certainty score with a range of acceptable certainty score values.
8. The method according to any one of embodiments 4 to 7, wherein the certainty measure relates to how uniform the sub-scan texture information is at the point, and/or to how close the texture data is to the known texture value of the determined tooth shade value, and/or to the amount of texture information used to derive the texture data at the point.
9. The method according to any one of embodiments 4 to 8, wherein the visual representation of the certainty score comprises a binary code, a bar structure with a color gradient, a numerical value, and/or a comparison between the texture data and the known texture value of the determined tooth shade value.
10. The method according to any one of the preceding embodiments, wherein the one or more reference tooth shade values relate to shade values for natural teeth with intact surface and/or to shade values for teeth prepared for a dental restoration.
11. The method according to any one of the preceding embodiments, wherein the method comprises comparing the texture data with known texture values for soft oral tissue, such as gum tissue and gingiva.
12. The method according to any of the previous embodiments, wherein the texture information comprises at least one of tooth color or surface roughness.
13. The method according to any one of the preceding embodiments, wherein the method comprises creating a shade profile for the tooth from shade values determined one or more points on the tooth.
14. The method according to embodiment 13, wherein the tooth shade profile comprises a one or more tooth shade regions on the tooth surface where an average tooth shade is derived for each region from tooth shade values determined for a number of points within the region.
15. The method according to any of the previous embodiments, wherein obtaining the digital 3D representation of the tooth comprises recording a series of sub-scans of the tooth, where at least one of said sub-scans comprises both texture information and geometry information for said tooth, and generating the digital 3D representation of the tooth from the recorded series of sub-scans.
16. The method according to embodiment 15, wherein the texture data at least partly are derived by combining the texture information from corresponding parts of a number of the sub-scans.
17. The method according to embodiment 16, wherein combining the texture information from the sub-scans comprises interpolating the texture information and/or calculating an average value of the texture information.
18. The method according to embodiment 17, wherein the calculated average value is a weighted average of the texture information.
19. A user interface for determining and displaying shade of a patient's tooth, wherein the user interface is configured for:
    • obtaining a digital 3D representation of the tooth, said digital 3D representation comprising shape data and texture data for the tooth;
    • displaying at least the shape data of the digital 3D representation such that the shape of the tooth is visualized in the user interface;
    • determining a tooth shade value for at least one point on the tooth based on the texture data of the corresponding point of the digital 3D representation and on known texture values of one or more reference tooth shade values; and
      displaying the determined tooth shade value.

Claims (7)

The invention claimed is:
1. A method for designing a denture, wherein the method comprises:
obtaining a digital 3D representation of the patient's soft tissue, said digital 3D representation comprising:
shape data expressing the shape of the soft tissue, and
texture data expressing a texture profile of the soft tissue; and
designing the denture such that a color of a soft tissue part of the denture is selected based on the texture profile of the corresponding part of the digital 3D representation,
wherein the step of obtaining the digital 3D representation is provided by recording a series of sub-scans that comprise the shape data and the texture data,
wherein the texture data at least partly is derived by combining texture information from corresponding parts of a number of the sub-scans.
2. The method according to claim 1, wherein the denture is an aesthetically pleasing denture defined by having the color that matches the color of the patient's soft tissue.
3. The method according to claim 1, wherein combining the texture information from the sub-scans comprises interpolating the texture information.
4. The method according to claim 1, wherein combining the texture information from the sub-scans comprises calculating an average value of the texture information.
5. The method according to claim 4, wherein the calculated average value is a weighted average of the texture information.
6. The method according to claim 1, wherein the method further comprises comparing the texture data with known texture values for soft tissue.
7. The method according to claim 1, wherein the method further comprises designing the denture such that a shape of the soft tissue part of the denture is selected based on the shape data of the corresponding part of the digital 3D representation.
US16/946,186 2014-02-07 2020-06-09 Detecting tooth shade Active 2035-05-05 US11701208B2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US16/946,186 US11701208B2 (en) 2014-02-07 2020-06-09 Detecting tooth shade
US17/742,813 US11723759B2 (en) 2014-02-07 2022-05-12 Detecting tooth shade
US17/742,955 US11707347B2 (en) 2014-02-07 2022-05-12 Detecting tooth shade

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
DKPA201470066 2014-02-07
DKPA201470066 2014-02-07
PCT/EP2015/052537 WO2015118120A1 (en) 2014-02-07 2015-02-06 Detecting tooth shade
US201615117078A 2016-08-05 2016-08-05
US15/888,764 US10695151B2 (en) 2014-02-07 2018-02-05 Detecting tooth shade
US16/946,186 US11701208B2 (en) 2014-02-07 2020-06-09 Detecting tooth shade

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US15/888,764 Continuation US10695151B2 (en) 2014-02-07 2018-02-05 Detecting tooth shade

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US17/742,813 Continuation US11723759B2 (en) 2014-02-07 2022-05-12 Detecting tooth shade
US17/742,955 Continuation US11707347B2 (en) 2014-02-07 2022-05-12 Detecting tooth shade

Publications (2)

Publication Number Publication Date
US20200352688A1 US20200352688A1 (en) 2020-11-12
US11701208B2 true US11701208B2 (en) 2023-07-18

Family

ID=52473894

Family Applications (5)

Application Number Title Priority Date Filing Date
US15/117,078 Active US10010387B2 (en) 2014-02-07 2015-02-06 Detecting tooth shade
US15/888,764 Active 2035-07-24 US10695151B2 (en) 2014-02-07 2018-02-05 Detecting tooth shade
US16/946,186 Active 2035-05-05 US11701208B2 (en) 2014-02-07 2020-06-09 Detecting tooth shade
US17/742,955 Active US11707347B2 (en) 2014-02-07 2022-05-12 Detecting tooth shade
US17/742,813 Active US11723759B2 (en) 2014-02-07 2022-05-12 Detecting tooth shade

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US15/117,078 Active US10010387B2 (en) 2014-02-07 2015-02-06 Detecting tooth shade
US15/888,764 Active 2035-07-24 US10695151B2 (en) 2014-02-07 2018-02-05 Detecting tooth shade

Family Applications After (2)

Application Number Title Priority Date Filing Date
US17/742,955 Active US11707347B2 (en) 2014-02-07 2022-05-12 Detecting tooth shade
US17/742,813 Active US11723759B2 (en) 2014-02-07 2022-05-12 Detecting tooth shade

Country Status (2)

Country Link
US (5) US10010387B2 (en)
WO (1) WO2015118120A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11986325B2 (en) * 2019-06-13 2024-05-21 Osstemimplant Co., Ltd. Treatment information display device and method for displaying treatment history on image of teeth in accumulated manner

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2763826C (en) 2009-06-17 2020-04-07 3Shape A/S Focus scanning apparatus
WO2015118120A1 (en) 2014-02-07 2015-08-13 3Shape A/S Detecting tooth shade
EP3018461A1 (en) * 2014-11-07 2016-05-11 3M Innovative Properties Company A method of making a dental restoration
US20170020636A1 (en) * 2015-04-16 2017-01-26 Hadi Akeel System and method for robotic digital scanning of teeth
EP3372193B1 (en) * 2017-03-08 2021-04-21 Ivoclar Vivadent AG Method for determining a material colour of a dental restoration
WO2019158442A1 (en) 2018-02-16 2019-08-22 3Shape A/S Intraoral scanning with surface differentiation
RU2020135294A (en) * 2018-03-28 2022-04-28 Конинклейке Филипс Н.В. METHOD AND SYSTEM FOR ASSESSING TEETH SHADES IN UNMANAGED ENVIRONMENT
ES2943138T3 (en) * 2018-05-18 2023-06-09 Dental Imaging Technologies Corp Angle-based Shade Matching Dental 3D Scanner
US10753734B2 (en) 2018-06-08 2020-08-25 Dentsply Sirona Inc. Device, method and system for generating dynamic projection patterns in a confocal camera
ES2891031T3 (en) * 2018-10-25 2022-01-25 Ivoclar Vivadent Ag Procedure for determining a tooth color
KR20200075623A (en) * 2018-12-18 2020-06-26 (주)제노레이 Dental Treatment Planning Apparatus Using Matching Of 2D Medical Image And 3D Medical Image And Method Thereof
WO2020185733A1 (en) * 2019-03-10 2020-09-17 Carestream Dental Llc Dental shade matching for multiple anatomical regions
EP3763324B1 (en) * 2019-07-09 2023-09-06 VITA-ZAHNFABRIK H. Rauter GmbH & Co. KG Support system for producing dental restorations and dental restoration system
EP3984495B1 (en) * 2020-10-13 2024-08-21 Ivoclar Vivadent AG Method for forming a dental restoration
US11191620B1 (en) 2021-06-03 2021-12-07 Oxilio Ltd Systems and methods for generating an augmented 3D digital model of an anatomical structure of a subject
US20230165666A1 (en) * 2021-12-01 2023-06-01 Dentsply Sirona Inc. Proposing dental restoration material parameters

Citations (308)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3878905A (en) 1972-04-07 1975-04-22 Hawera Probst Kg Hartmetall Drill, especially rock drill
US3971065A (en) 1975-03-05 1976-07-20 Eastman Kodak Company Color imaging array
US4291958A (en) 1980-10-03 1981-09-29 Eastman Kodak Company Camera with electronic flash and piezoelectric lens motor
US4342227A (en) 1980-12-24 1982-08-03 International Business Machines Corporation Planar semiconductor three direction acceleration detecting device and method of fabrication
US4349880A (en) 1979-03-19 1982-09-14 Rca Corporation Inspection system for detecting defects in regular patterns
US4516231A (en) 1982-08-26 1985-05-07 Rca Corporation Optical disc system having momentum compensation
US4575805A (en) 1980-12-24 1986-03-11 Moermann Werner H Method and apparatus for the fabrication of custom-shaped implants
US4629324A (en) 1983-12-29 1986-12-16 Robotic Vision Systems, Inc. Arrangement for measuring depth based on lens focusing
US4640620A (en) 1983-12-29 1987-02-03 Robotic Vision Systems, Inc. Arrangement for rapid depth measurement using lens focusing
JPS62100716A (en) 1985-10-29 1987-05-11 Matsushita Electric Ind Co Ltd Photographing device
WO1988007695A1 (en) 1987-03-27 1988-10-06 The Board Of Trustees Of The Leland Stanford Junio Scanning confocal optical microscope
US4781448A (en) 1987-03-02 1988-11-01 Medical Concepts Inc. Zoom lens adapter for endoscopic camera
US4802759A (en) 1986-08-11 1989-02-07 Goro Matsumoto Three-dimensional shape measuring apparatus
US4896015A (en) 1988-07-29 1990-01-23 Refractive Laser Research & Development Program, Ltd. Laser delivery system
US5131844A (en) 1991-04-08 1992-07-21 Foster-Miller, Inc. Contact digitizer, particularly for dental applications
WO1992014118A1 (en) 1991-02-12 1992-08-20 Oxford Sensor Technology Limited An optical sensor
WO1992015034A1 (en) 1991-02-19 1992-09-03 Phoenix Laser Systems, Inc. Target movement compensation for laser surgical system
US5151609A (en) 1989-08-02 1992-09-29 Hitachi, Ltd. Method of detecting solid shape of object with autofocusing and image detection at each focus level
US5181181A (en) 1990-09-27 1993-01-19 Triton Technologies, Inc. Computer apparatus input device for three-dimensional information
US5269325A (en) 1989-05-26 1993-12-14 Biomagnetic Technologies, Inc. Analysis of biological signals using data from arrays of sensors
JPH06201337A (en) 1992-12-28 1994-07-19 Hitachi Ltd Surface shape detection device and its method
US5339154A (en) 1991-10-15 1994-08-16 Kaltenbach & Voigt Gmb & Co. Method and apparatus for optical measurement of objects
US5372502A (en) 1988-09-02 1994-12-13 Kaltenbach & Voight Gmbh & Co. Optical probe and method for the three-dimensional surveying of teeth
US5377011A (en) 1991-09-06 1994-12-27 Koch; Stephen K. Scanning system for three-dimensional object digitizing
US5428450A (en) 1992-12-22 1995-06-27 Bertin & Cie Method and apparatus for determining the color of an object that is transparent, diffusing, and absorbent, such as a tooth, in particular
US5455899A (en) 1992-12-31 1995-10-03 International Business Machines Corporation High speed image data processing circuit
US5563343A (en) 1993-05-26 1996-10-08 Cornell Research Foundation, Inc. Microelectromechanical lateral accelerometer
DE19524855A1 (en) 1995-07-07 1997-01-09 Siemens Ag Method and device for computer-aided restoration of teeth
US5605459A (en) 1995-04-14 1997-02-25 Unisn Incorporated Method of and apparatus for making a dental set-up model
US5615003A (en) 1994-11-29 1997-03-25 Hermary; Alexander T. Electromagnetic profile scanner
WO1997014932A1 (en) 1995-10-20 1997-04-24 Optronic Consult Ab Process and device for the measuring of a three-dimensional shape
US5675407A (en) 1995-03-02 1997-10-07 Zheng Jason Geng Color ranging method for high speed low-cost three dimensional surface profile measurement
US5702249A (en) 1995-05-19 1997-12-30 Cooper; David H. Modular intra-oral imaging system video camera
DE19642247C1 (en) 1996-10-12 1998-01-15 Sebastian Meller Prosthesis production system for prepared tooth
US5722412A (en) 1996-06-28 1998-03-03 Advanced Technology Laboratories, Inc. Hand held ultrasonic diagnostic instrument
US5737339A (en) 1994-02-08 1998-04-07 Fujitsu Limited Method and apparatus for verifying stored data
US5737084A (en) 1995-09-29 1998-04-07 Takaoka Electric Mtg. Co., Ltd. Three-dimensional shape measuring apparatus
US5766006A (en) 1995-06-26 1998-06-16 Murljacic; Maryann Lehmann Tooth shade analyzer system and methods
WO1998045745A1 (en) 1997-04-04 1998-10-15 Isis Innovation Limited Microscopy imaging apparatus and method
US5850289A (en) 1994-08-24 1998-12-15 Tricorder Technology Plc Scanning arrangement and method
US5851113A (en) 1996-01-02 1998-12-22 Lj Laboratories, L.L.C. Apparatus and method for measuring the color of teeth
WO1999047964A1 (en) 1998-03-20 1999-09-23 Genetic Microsystems, Inc. Wide field of view and high speed scanning microscopy
US6007332A (en) 1996-09-26 1999-12-28 O'brien; William J. Tooth color matching system
US6026189A (en) 1997-11-13 2000-02-15 National Research Council Of Canada Method of recognizing objects within two-dimensional and three-dimensional images
WO2000008415A1 (en) 1998-08-05 2000-02-17 Cadent Ltd. Imaging a three-dimensional structure by confocal focussing an array of light beams
US6072496A (en) 1998-06-08 2000-06-06 Microsoft Corporation Method and system for capturing and representing 3D geometry, color and shading of facial expressions and other animated objects
US6081739A (en) 1998-05-21 2000-06-27 Lemchen; Marc S. Scanning device or methodology to produce an image incorporating correlated superficial, three dimensional surface and x-ray images and measurements of an object
US6135961A (en) 1996-06-28 2000-10-24 Sonosite, Inc. Ultrasonic signal processor for a hand held ultrasonic diagnostic instrument
US6148120A (en) 1997-10-30 2000-11-14 Cognex Corporation Warping of focal images to correct correspondence error
WO2001011193A1 (en) 1999-08-05 2001-02-15 Jama Mining Equipment Ab Arrangement for fixing rock bolts during rock reinforcement
US6206691B1 (en) 1998-05-20 2001-03-27 Shade Analyzing Technologies, Inc. System and methods for analyzing tooth shades
US6229913B1 (en) 1995-06-07 2001-05-08 The Trustees Of Columbia University In The City Of New York Apparatus and methods for determining the three-dimensional shape of an object using active illumination and relative blurring in two-images due to defocus
US6227850B1 (en) 1999-05-13 2001-05-08 Align Technology, Inc. Teeth viewing system
US6249616B1 (en) 1997-05-30 2001-06-19 Enroute, Inc Combining digital images based on three-dimensional relationships between source image data sets
US6251073B1 (en) 1999-08-20 2001-06-26 Novasonics, Inc. Miniaturized ultrasound apparatus and method
US6259452B1 (en) 1997-04-14 2001-07-10 Massachusetts Institute Of Technology Image drawing system and method with real-time occlusion culling
US6263234B1 (en) 1996-10-01 2001-07-17 Leica Microsystems Heidelberg Gmbh Confocal surface-measuring device
US20010030748A1 (en) 1998-07-09 2001-10-18 Jung Wayne D. Apparatus and method for measuring optical characteristics of an object
WO2001084479A1 (en) 2000-04-28 2001-11-08 Orametirix, Inc. Method and system for scanning a surface and generating a three-dimensional object
US6334853B1 (en) 1997-05-22 2002-01-01 Cadent Ltd Method for obtaining a dental occlusion map
US6334773B1 (en) 1997-02-24 2002-01-01 Dentronic Ab Method and arrangement for making artificial teeth
US6361489B1 (en) 1998-11-25 2002-03-26 Jory Tsai Medical inspection device
US6450807B1 (en) 1997-06-20 2002-09-17 Align Technology, Inc. System and method for positioning teeth
WO2002076327A1 (en) 2001-03-23 2002-10-03 Decim Ab A method of and an arrangement for a dental restoration
US6471511B1 (en) 1997-06-20 2002-10-29 Align Technology, Inc. Defining tooth-moving appliances computationally
US6476803B1 (en) 2000-01-06 2002-11-05 Microsoft Corporation Object modeling system and process employing noise elimination and robust surface extraction techniques
US6485413B1 (en) 1991-04-29 2002-11-26 The General Hospital Corporation Methods and apparatus for forward-directed optical scanning instruments
US20030043089A1 (en) 2001-08-17 2003-03-06 Eric Hanson Doubling of speed in CMOS sensor with column-parallel ADCS
US6532299B1 (en) 2000-04-28 2003-03-11 Orametrix, Inc. System and method for mapping a surface
US20030096210A1 (en) 1999-11-30 2003-05-22 Orametrix, Inc. Interactive orthodontic care system based on intra-oral scanning of teeth
US6575751B1 (en) 1998-11-03 2003-06-10 Shade Analyzing Technologies, Inc. Interactive dental restorative network
US6592371B2 (en) 2000-10-25 2003-07-15 Duane Durbin Method and system for imaging and modeling a three dimensional structure
WO2003060587A1 (en) 2002-01-15 2003-07-24 École Polytechnique Fédérale de Lausanne Microscopy imaging apparatus and mehod for generating an image
US20030156283A1 (en) 2002-02-21 2003-08-21 Lj Laboratories, L.L.C. Miniaturized system and method for measuring optical characteristics
US20030158482A1 (en) 2002-02-20 2003-08-21 Koninklijke Philips Electronics N.V. Portable 3D ultrasound system
US20030164952A1 (en) 2000-08-25 2003-09-04 Nikolaj Deichmann Method and apparatus for three-dimensional optical scanning of interior surfaces
US6645148B2 (en) 2001-03-20 2003-11-11 Vermon Ultrasonic probe including pointing devices for remotely controlling functions of an associated imaging system
JP2004029685A (en) 2002-06-23 2004-01-29 Akira Ishii Method and device for fixed magnification imaging using variable focal lens
US20040080754A1 (en) 2002-10-29 2004-04-29 Mitutoyo Corporation Interferometer using integrated imaging array and high-density phase-shifting array
US6743014B2 (en) 2000-05-16 2004-06-01 Ivoclar Vivadent Ag Shade determination apparatus and method for specifying and determining colors for teeth and dental restorations
US6751344B1 (en) 1999-05-28 2004-06-15 Champion Orthotic Investments, Inc. Enhanced projector system for machine vision
US6750873B1 (en) 2000-06-27 2004-06-15 International Business Machines Corporation High quality texture reconstruction from multiple scans
US20040125103A1 (en) 2000-02-25 2004-07-01 Kaufman Arie E. Apparatus and method for volume processing and rendering
US6761561B2 (en) 2002-06-07 2004-07-13 Schick Technologies Wireless dental camera
WO2004066615A1 (en) 2003-01-22 2004-08-05 Nokia Corporation Image control
US20040155975A1 (en) 2002-09-17 2004-08-12 Hart Douglas P. 3-D imaging system
US20040185422A1 (en) 2003-03-21 2004-09-23 Sirona Dental Systems Gmbh Data base, tooth model and restorative item constructed from digitized images of real teeth
US20040204787A1 (en) 2003-04-03 2004-10-14 Avi Kopelman Method and system for fabricating a dental coping, and a coping fabricated thereby
DE10321883A1 (en) 2003-05-07 2004-12-09 Universität Stuttgart Triangulation measurement device for determining object 3D structure has illumination and observation arrays with a projected pattern being evaluated using cross correlation or phase shift analysis
US20040254476A1 (en) 2003-03-24 2004-12-16 Henley Quadling Laser digitizer system for dental applications
US20050020910A1 (en) 2003-04-30 2005-01-27 Henley Quadling Intra-oral imaging system
US6865289B1 (en) 2000-02-07 2005-03-08 Canon Kabushiki Kaisha Detection and removal of image occlusion errors
US20050057745A1 (en) 2003-09-17 2005-03-17 Bontje Douglas A. Measurement methods and apparatus
US20050074718A1 (en) 2003-10-03 2005-04-07 Pye Graham Tooth shade scan system and method
JP2005098833A (en) 2003-09-25 2005-04-14 Keyence Corp Displacement meter and displacement measuring method
US20050090749A1 (en) 1996-09-02 2005-04-28 Rudger Rubbert Method and device for carrying out optical pick up
US6904159B2 (en) 2001-12-20 2005-06-07 Mitsubishi Electric Research Laboratories, Inc. Identifying moving objects in a video using volume growing and change detection masks
US20050142517A1 (en) 2003-12-30 2005-06-30 Howard Frysh System for producing a dental implant and method
WO2005067389A2 (en) 2004-01-15 2005-07-28 Technion Research & Development Foundation Ltd. Three-dimensional video scanner
US20050212756A1 (en) 2004-03-23 2005-09-29 Marvit David L Gesture based navigation of a handheld user interface
US20050212753A1 (en) 2004-03-23 2005-09-29 Marvit David L Motion controlled remote controller
US6954550B2 (en) 2000-11-29 2005-10-11 Omron Corporation Image processing method and apparatus
US20050232509A1 (en) 2004-04-16 2005-10-20 Andrew Blake Virtual image artifact detection
US20050237581A1 (en) 2004-04-21 2005-10-27 Knighton Mark S Hand held portable three dimensional scanner
US20050243330A1 (en) 2004-04-28 2005-11-03 Simon Magarill Methods and apparatus for determining three dimensional configurations
US6967644B1 (en) 1998-10-01 2005-11-22 Canon Kabushiki Kaisha Coordinate input apparatus and control method thereof, and computer readable memory
US6975898B2 (en) 2000-06-19 2005-12-13 University Of Washington Medical imaging, diagnosis, and therapy using a scanning single optical fiber system
US6977732B2 (en) 2002-12-26 2005-12-20 National Taiwan University Miniature three-dimensional contour scanner
US20050283065A1 (en) 2004-06-17 2005-12-22 Noam Babayoff Method for providing data associated with the intraoral cavity
US6990228B1 (en) 1999-12-17 2006-01-24 Canon Kabushiki Kaisha Image processing apparatus
US20060020204A1 (en) 2004-07-01 2006-01-26 Bracco Imaging, S.P.A. System and method for three-dimensional space management and visualization of ultrasound data ("SonoDEX")
US20060025684A1 (en) 2001-04-19 2006-02-02 Sonosite, Inc. Medical diagnostic ultrasound instrument with ECG module, authorization mechanism and methods of use
US7010223B2 (en) 2001-05-26 2006-03-07 Dürr Dental GmbH & Co. KG Dental or endoscopic camera
US20060072189A1 (en) 2004-10-06 2006-04-06 Dimarzio Charles A Confocal reflectance microscope system with dual rotating wedge scanner assembly
US20060072123A1 (en) 2003-01-25 2006-04-06 Wilson John E Methods and apparatus for making images including depth information
US7027642B2 (en) 2000-04-28 2006-04-11 Orametrix, Inc. Methods for registration of three-dimensional frames to create three-dimensional virtual models of objects
US20060092133A1 (en) 2004-11-02 2006-05-04 Pierre A. Touma 3D mouse and game controller based on spherical coordinates system and system for use
US7058213B2 (en) 1999-03-08 2006-06-06 Orametrix, Inc. Scanning system and calibration method for capturing precise three-dimensional information of objects
US20060127852A1 (en) 2004-12-14 2006-06-15 Huafeng Wen Image based orthodontic treatment viewing system
WO2006065955A2 (en) 2004-12-14 2006-06-22 Orthoclear Holdings, Inc. Image based orthodontic treatment methods
US7077647B2 (en) 2002-08-22 2006-07-18 Align Technology, Inc. Systems and methods for treatment analysis by teeth matching
US7079679B2 (en) 2000-09-27 2006-07-18 Canon Kabushiki Kaisha Image processing apparatus
US7086863B2 (en) 2001-04-23 2006-08-08 Cicero Dental Systems, B.V. Method for production of an artificial tooth
US7099732B2 (en) 1999-03-29 2006-08-29 Genex Technologies, Inc. Sanitary sleeve or tip for intra-oral three-dimensional camera
US20060212260A1 (en) 2005-03-03 2006-09-21 Cadent Ltd. System and method for scanning an intraoral cavity
US7123760B2 (en) 2002-11-21 2006-10-17 General Electric Company Method and apparatus for removing obstructing structures in CT imaging
US20060251408A1 (en) 2004-01-23 2006-11-09 Olympus Corporation Image processing system and camera
US7134874B2 (en) 1997-06-20 2006-11-14 Align Technology, Inc. Computer automated development of an orthodontic treatment plan and appliance
US7160110B2 (en) 1999-11-30 2007-01-09 Orametrix, Inc. Three-dimensional occlusal and interproximal contact detection and display using virtual tooth models
US20070015025A1 (en) 2005-07-15 2007-01-18 Honda Motor Co., Ltd. Membrane electrode assembly for solid polymer electrolyte fuel cell
US7166537B2 (en) 2002-03-18 2007-01-23 Sarcos Investments Lc Miniaturized imaging device with integrated circuit connector system
CN1906678A (en) 2004-04-21 2007-01-31 松下电器产业株式会社 Confocal optical system aperture position detector, confocal optical system aperture position controller, optical head, optical information processor and confocal optical system aperture position dete
US20070026363A1 (en) 1998-11-03 2007-02-01 Maryann Lehmann Interactive dental restorative network
US20070031774A1 (en) 2005-08-03 2007-02-08 3M Innovative Properties Company Registering physical and virtual tooth structures with markers
US20070041729A1 (en) 2002-10-23 2007-02-22 Philip Heinz Systems and methods for detecting changes in incident optical radiation at high frequencies
CN1934481A (en) 2004-03-19 2007-03-21 塞隆纳牙科系统有限责任公司 High-speed measuring device and method based on a confocal microscopy principle
US20070064242A1 (en) 2005-09-01 2007-03-22 Sonaray, Inc. Method and system for obtaining high resolution 3-D images of moving objects by use of sensor fusion
JP2007072103A (en) 2005-09-06 2007-03-22 Fujifilm Holdings Corp Camera
US20070078340A1 (en) 2005-09-30 2007-04-05 Siemens Medical Solutions Usa, Inc. Method and apparatus for controlling ultrasound imaging systems having positionable transducers
US7213214B2 (en) 2001-06-12 2007-05-01 Idelix Software Inc. Graphical user interface with zoom for detail-in-context presentations
US7215430B2 (en) 1996-04-24 2007-05-08 Leica Geosystems Hds Llc Integrated system for quickly and accurately imaging and modeling three-dimensional objects
US20070103460A1 (en) 2005-11-09 2007-05-10 Tong Zhang Determining camera motion
US7221332B2 (en) 2003-12-19 2007-05-22 Eastman Kodak Company 3D stereo OLED display
US7230771B2 (en) 2002-10-25 2007-06-12 Koninklijke Philips Electronics N.V. Zoom lens
US20070134615A1 (en) 2005-12-08 2007-06-14 Lovely Peter S Infrared dental imaging
US20070140539A1 (en) 2005-12-19 2007-06-21 Olympus Corporation Image combining apparatus
US20070171220A1 (en) 2006-01-20 2007-07-26 Kriveshko Ilya A Three-dimensional scan recovery
WO2007084727A1 (en) 2006-01-20 2007-07-26 3M Innovative Properties Company Digital dentistry
US20070182812A1 (en) 2004-05-19 2007-08-09 Ritchey Kurtis J Panoramic image-based virtual reality/telepresence audio-visual system and method
US20070212667A1 (en) 2006-03-13 2007-09-13 Jung Wayne D Systems and methods for preparing dental restorations
US20070252074A1 (en) 2004-10-01 2007-11-01 The Board Of Trustees Of The Leland Stanford Junio Imaging Arrangements and Methods Therefor
US7296996B2 (en) 1999-11-30 2007-11-20 Orametrix, Inc. Virtual bracket placement and evaluation
US7339170B2 (en) 2003-07-16 2008-03-04 Shrenik Deliwala Optical encoding and reconstruction
US20080058783A1 (en) 2003-11-04 2008-03-06 Palomar Medical Technologies, Inc. Handheld Photocosmetic Device
US20080063998A1 (en) 2006-09-12 2008-03-13 Rongguang Liang Apparatus for caries detection
US20080071143A1 (en) 2006-09-18 2008-03-20 Abhishek Gattani Multi-dimensional navigation of endoscopic video
US20080070684A1 (en) 2006-09-14 2008-03-20 Mark Haigh-Hutchinson Method and apparatus for using a common pointing input to control 3D viewpoint and object targeting
US7349104B2 (en) 2003-10-23 2008-03-25 Technest Holdings, Inc. System and a method for three-dimensional imaging systems
US7355721B2 (en) 2003-05-05 2008-04-08 D4D Technologies, Llc Optical coherence tomography imaging
US20080118886A1 (en) 2006-11-21 2008-05-22 Rongguang Liang Apparatus for dental oct imaging
US20080132886A1 (en) 2004-04-09 2008-06-05 Palomar Medical Technologies, Inc. Use of fractional emr technology on incisions and internal tissues
US20080131028A1 (en) 2006-11-30 2008-06-05 Pillman Bruce H Producing low resolution images
US7385708B2 (en) 2002-06-07 2008-06-10 The University Of North Carolina At Chapel Hill Methods and systems for laser based real-time structured light depth extraction
DE102007005726A1 (en) 2007-01-31 2008-08-07 Sirona Dental Systems Gmbh Device and method for 3D optical measurement
US20080194928A1 (en) 2007-01-05 2008-08-14 Jadran Bandic System, device, and method for dermal imaging
US20080194950A1 (en) 2007-02-13 2008-08-14 General Electric Company Ultrasound imaging remote control unit
JP2008194108A (en) 2007-02-09 2008-08-28 Shiyoufuu:Kk Three-dimensional characteristic measuring and displaying apparatus with positional direction detecting function
WO2008125605A2 (en) 2007-04-13 2008-10-23 Michael Schwertner Method and assembly for optical reproduction with depth discrimination
US7460248B2 (en) 2006-05-15 2008-12-02 Carestream Health, Inc. Tissue imaging system
US7471821B2 (en) 2000-04-28 2008-12-30 Orametrix, Inc. Method and apparatus for registering a known digital object to scanned 3-D model
US7474414B2 (en) 2005-06-01 2009-01-06 Inus Technology, Inc. System and method of guiding real-time inspection using 3D scanners
US7483062B2 (en) 2004-04-28 2009-01-27 International Business Machines Corporation Method for removal of moving objects from a video stream
US20090040175A1 (en) 2004-12-22 2009-02-12 Rex Fang Xu Input interface device with transformable form factor
US7494338B2 (en) 2005-01-11 2009-02-24 Duane Durbin 3D dental scanner
US20090061381A1 (en) 2007-09-05 2009-03-05 Duane Milford Durbin Systems and methods for 3D previewing
WO2009026645A1 (en) 2007-08-31 2009-03-05 Signostics Pty Ltd Apparatus and method for medical scanning
US20090076321A1 (en) 2005-12-26 2009-03-19 Kanagawa Furniture Co., Ltd. Digital camera for taking image inside oral cavity
WO2009034157A1 (en) 2007-09-12 2009-03-19 Degudent Gmbh Method for determining the position of an intraoral measuring device
US20090087050A1 (en) 2007-08-16 2009-04-02 Michael Gandyra Device for determining the 3D coordinates of an object, in particular of a tooth
US20090097108A1 (en) 2005-05-12 2009-04-16 Fox William J Confocal scanning microscope having optical and scanning systems which provide a handheld imaging head
US7522322B2 (en) 2004-08-19 2009-04-21 Carestream Health, Inc. Apparatus for dental shade measurement
US20090103103A1 (en) 2007-10-18 2009-04-23 Mht Optic Research Ag Device for tomographic scanning objects
CN101426085A (en) 2004-10-01 2009-05-06 科兰·斯坦福青年大学托管委员会 Imaging arrangements and methods therefor
WO2009063088A2 (en) 2007-11-15 2009-05-22 Sirona Dental Systems Gmbh Method for optical measurement of objects using a triangulation method
US20090133260A1 (en) 2007-11-26 2009-05-28 Ios Technologies, Inc 3D dental shade matching and apparatus
US7550707B2 (en) 2007-03-01 2009-06-23 Sony Corporation Biometrics authentication system with authentication based upon multiple layers
US7551353B2 (en) 2005-07-27 2009-06-23 Samsung Electronics Co., Ltd. Glassless stereoscopic display
US20090160858A1 (en) 2007-12-21 2009-06-25 Industrial Technology Research Institute Method for reconstructing three dimensional model
US20090167948A1 (en) 2005-02-08 2009-07-02 Berman Steven T System and Method for Selective Image Capture, Transmission and Reconstruction
US20090177050A1 (en) 2006-07-17 2009-07-09 Medrad, Inc. Integrated medical imaging systems
WO2009089126A1 (en) 2008-01-04 2009-07-16 3M Innovative Properties Company Three-dimensional model refinement
CN101513350A (en) 2008-02-22 2009-08-26 西门子公司 Device and method for displaying medical image and imaging system
US20090231649A1 (en) 2008-03-12 2009-09-17 Sirat Gabriel Y Intraoral Imaging System and Method based on Conoscopic Holography
US20090233253A1 (en) 2007-12-21 2009-09-17 Mrazek William R Dental shade guide
JP2009238245A (en) 2009-07-13 2009-10-15 Namco Bandai Games Inc Image generation system and information storage medium
US7609875B2 (en) 2005-05-27 2009-10-27 Orametrix, Inc. Scanner system and method for mapping surface of three-dimensional object
US7636455B2 (en) 2002-06-04 2009-12-22 Raytheon Company Digital image edge detection and road network tracking method and system
US20090322676A1 (en) 2007-09-07 2009-12-31 Apple Inc. Gui applications for use with 3d remote controller
US20100009308A1 (en) 2006-05-05 2010-01-14 Align Technology, Inc. Visualizing and Manipulating Digital Models for Dental Treatment
US20100079581A1 (en) 2008-09-30 2010-04-01 Texas Instruments Incorporated 3d camera using flash with structured light
US20100085636A1 (en) 2008-10-06 2010-04-08 Mht Optic Research Ag Optical System for a Confocal Microscope
US7708557B2 (en) 2006-10-16 2010-05-04 Natural Dental Implants Ag Customized dental prosthesis for periodontal- or osseointegration, and related systems and methods
WO2010064156A1 (en) 2008-12-03 2010-06-10 Koninklijke Philips Electronics, N.V. Ultrasound assembly and system comprising interchangable transducers and displays
EP2200332A1 (en) 2008-12-17 2010-06-23 Robert Bosch GmbH Autostereoscopic display
US20100157086A1 (en) 2008-12-15 2010-06-24 Illumina, Inc Dynamic autofocus method and system for assay imager
US20100156901A1 (en) 2008-12-22 2010-06-24 Electronics And Telecommunications Research Institute Method and apparatus for reconstructing 3d model
US7762814B2 (en) 2004-09-14 2010-07-27 Oratio B.V. Method of manufacturing and installing a ceramic dental implant with an aesthetic implant abutment
US20100201986A1 (en) 2004-02-20 2010-08-12 Jean-Marc Inglese Equipment and method for measuring dental shade
US20100231509A1 (en) 2009-03-12 2010-09-16 Marc Boillot Sterile Networked Interface for Medical Systems
WO2010106379A1 (en) 2009-03-20 2010-09-23 Mediwatch Uk Limited Ultrasound probe with accelerometer
US20100239136A1 (en) 2008-09-18 2010-09-23 Steinbichler Optotechnik Gmbh Device for Determining the 3D Coordinates of an Object, In Particular of a Tooth
US20100268069A1 (en) 2009-04-16 2010-10-21 Rongguang Liang Dental surface imaging using polarized fringe projection
US7831292B2 (en) 2002-03-06 2010-11-09 Mako Surgical Corp. Guidance system and method for surgical procedures with improved feedback
DE102009023952A1 (en) 2009-06-04 2010-12-09 DüRR DENTAL AG Method for determining tooth color of e.g. dental prosthesis, involves providing difference between color values of images and tooth color pattern as measurement for correlating colors of tooth region and tooth color pattern
WO2010145669A1 (en) 2009-06-17 2010-12-23 3Shape A/S Focus scanning apparatus
WO2011011193A1 (en) 2009-07-21 2011-01-27 Dimensional Photonics International, Inc. Integrated display in a hand-held three-dimensional metrology system
US7929751B2 (en) 2005-11-09 2011-04-19 Gi, Llc Method and apparatus for absolute-coordinate three-dimensional surface imaging
US7929151B2 (en) 2008-01-11 2011-04-19 Carestream Health, Inc. Intra-oral camera for diagnostic and cosmetic imaging
WO2011047731A1 (en) 2009-10-22 2011-04-28 Tele Atlas B.V. Method for creating a mosaic image using masks
EP2325771A2 (en) 2009-11-24 2011-05-25 Sirona Dental Systems GmbH Systems, methods, apparatuses, and computer-readable storage media for designing and manufacturing prosthetic dental items
US20110188726A1 (en) 2008-06-18 2011-08-04 Ram Nathaniel Method and system for stitching multiple images into a panoramic image
US20110200249A1 (en) 2010-02-17 2011-08-18 Harris Corporation Surface detection in images based on spatial data
US8003889B2 (en) 2007-08-02 2011-08-23 Thomas & Betts International, Inc. Conduit sleeve pass through for concrete construction
WO2011120526A1 (en) 2010-03-30 2011-10-06 3Shape A/S Scanning of cavities with restricted accessibility
US8078006B1 (en) 2001-05-04 2011-12-13 Legend3D, Inc. Minimal artifact image sequence depth enhancement system and method
US20110310449A1 (en) 2010-06-17 2011-12-22 Eun-Soo Kim Method for generating 3d video computer-generated hologram using look-up table and temporal redundancy and apparatus thereof
US20110316978A1 (en) 2009-02-25 2011-12-29 Dimensional Photonics International, Inc. Intensity and color display for a three-dimensional metrology system
US8090194B2 (en) 2006-11-21 2012-01-03 Mantis Vision Ltd. 3D geometric modeling and motion capture using both single and dual imaging
WO2012000511A1 (en) 2010-06-29 2012-01-05 3Shape A/S 2d image arrangement
US8092215B2 (en) 2008-05-23 2012-01-10 Align Technology, Inc. Smile designer
WO2012007003A1 (en) 2010-07-12 2012-01-19 3Shape A/S 3D modeling of an object using textural features
US20120015316A1 (en) 2001-04-13 2012-01-19 Rohit Sachdeva Unified three dimensional virtual craniofacial and dentition model and uses thereof
US8103134B2 (en) 2008-02-20 2012-01-24 Samsung Electronics Co., Ltd. Method and a handheld device for capturing motion
US8121351B2 (en) 2008-03-09 2012-02-21 Microsoft International Holdings B.V. Identification of objects in a 3D video using non/over reflective clothing
US20120062557A1 (en) 2010-09-10 2012-03-15 Dimensional Photonics International, Inc. Systems and methods for processing and displaying intra-oral measurement data
US8144954B2 (en) 2007-11-08 2012-03-27 D4D Technologies, Llc Lighting compensated dynamic texture mapping of 3-D models
US8180100B2 (en) 2004-08-11 2012-05-15 Honda Motor Co., Ltd. Plane detector and detecting method
US8177551B2 (en) 2001-04-13 2012-05-15 Orametrix, Inc. Method and system for comprehensive evaluation of orthodontic treatment using unified workstation
US20120141949A1 (en) 2010-10-12 2012-06-07 Larry Bodony System and Apparatus for Haptically Enabled Three-Dimensional Scanning
WO2012076013A1 (en) 2010-12-06 2012-06-14 3Shape A/S System with 3d user interface integration
US8208704B2 (en) 2010-07-13 2012-06-26 Carestream Health, Inc. Dental shade mapping
WO2012083960A1 (en) 2010-12-22 2012-06-28 3Shape A/S System and method for scanning objects being modified
US20120179035A1 (en) 2011-01-07 2012-07-12 General Electric Company Medical device with motion sensing
US20120195471A1 (en) 2011-01-31 2012-08-02 Microsoft Corporation Moving Object Segmentation Using Depth Images
WO2012115862A2 (en) 2011-02-22 2012-08-30 3M Innovative Properties Company Space carving in 3d data acquisition
US8260539B2 (en) 2010-05-12 2012-09-04 GM Global Technology Operations LLC Object and vehicle detection and tracking using 3-D laser rangefinder
US8280152B2 (en) 2007-11-15 2012-10-02 Sirona Dental Systems Gmbh Method for optical measurement of the three dimensional geometry of objects
US8331653B2 (en) 2004-08-11 2012-12-11 Tokyo Institute Of Technology Object detector
US8335353B2 (en) 2007-04-04 2012-12-18 Sony Corporation Biometrics authentication system
US8345961B2 (en) 2008-09-10 2013-01-01 Huawei Device Co., Ltd. Image stitching method and apparatus
WO2013010910A1 (en) 2011-07-15 2013-01-24 3Shape A/S Detection of a movable object when 3d scanning a rigid object
US20130034823A1 (en) 2011-08-02 2013-02-07 Rongguang Liang Adaptive illumination method and apparatus for dental shade matching
US8384665B1 (en) 2006-07-14 2013-02-26 Ailive, Inc. Method and system for making a selection in 3D virtual environment
US8386061B2 (en) 2008-06-02 2013-02-26 Dentsply International Inc. Methods for designing a customized dental prosthesis using digital images of a patient
US8390821B2 (en) 2005-10-11 2013-03-05 Primesense Ltd. Three-dimensional sensing using speckle patterns
US20130110469A1 (en) 2010-07-19 2013-05-02 Avi Kopelman Methods and systems for creating and interacting with three dimensional virtual models
US8442283B2 (en) 2006-08-30 2013-05-14 Anatomage Inc. Patient-specific three-dimensional dentition model
US8456636B2 (en) 2007-10-01 2013-06-04 Ferton Holding, S.A. Device for detecting signs of bacterial infection of the teeth
US20130158694A1 (en) 2011-03-18 2013-06-20 Natural Dental Implants Ag Integrated Support Device For Providing Temporary Primary Stability to Dental Implants and Prosthesis, and Related Methods
US8469705B2 (en) 2001-04-13 2013-06-25 Orametrix, Inc. Method and system for integrated orthodontic treatment planning using unified workstation
US8477320B2 (en) 2009-05-15 2013-07-02 Degudent Gmbh Method and measuring arrangement for the three-dimensional measurement of an object
EP2620733A1 (en) 2012-01-27 2013-07-31 Ivoclar Vivadent AG Dental device
WO2013122662A1 (en) 2012-02-13 2013-08-22 3M Innovative Properties Company Dental milling block containing individualized dental article and process of production
US8526700B2 (en) 2010-10-06 2013-09-03 Robert E. Isaacs Imaging system and method for surgical and interventional medical procedures
US20130244197A1 (en) 2012-03-16 2013-09-19 Soek Gam Tjioe Dental Shade Matching Device
US8547374B1 (en) 2009-07-24 2013-10-01 Lockheed Martin Corporation Detection and reconstruction of 3D objects with passive imaging sensors
US20130260340A1 (en) 2012-03-29 2013-10-03 3M Innovative Properties Company Powder for enhancing feature contrast for intraoral digital image scanning
US8564657B2 (en) 2009-05-29 2013-10-22 Honda Research Institute Europe Gmbh Object motion detection system based on combining 3D warping techniques and a proper object motion detection
US8570530B2 (en) 2009-06-03 2013-10-29 Carestream Health, Inc. Apparatus for dental surface shape and shade imaging
US8571281B2 (en) 2010-07-13 2013-10-29 Carestream Health, Inc. Dental shade mapping
US8571397B2 (en) 2008-11-21 2013-10-29 Carestream Health, Inc. Auto focus intraoral camera with liquid lens
EP2664272A1 (en) 2011-01-11 2013-11-20 Kabushiki Kaisya Advance Oral imaging and display system
US8625854B2 (en) 2005-09-09 2014-01-07 Industrial Research Limited 3D scene scanner and a position and orientation system
US20140022356A1 (en) 2010-12-21 2014-01-23 3Shape A/S Optical system in 3d focus scanner
US20140022352A1 (en) 2010-12-21 2014-01-23 3Shape A/S Motion blur compensation
US20140146142A1 (en) 2011-07-08 2014-05-29 François Duret Three-dimensional measuring device used in the dental field
US8743114B2 (en) 2008-09-22 2014-06-03 Intel Corporation Methods and systems to determine conservative view cell occlusion
WO2014125037A1 (en) 2013-02-13 2014-08-21 3Shape A/S Focus scanning apparatus recording color
US8828287B2 (en) 2009-03-25 2014-09-09 Oratio B.V. Veneered dental restoration with a controlled shade
US20140255878A1 (en) 2013-03-07 2014-09-11 A.Tron3D Gmbh Method for optical acquisition of the three-dimensional geometry of objects
US8867820B2 (en) 2009-10-07 2014-10-21 Microsoft Corporation Systems and methods for removing a background of an image
EP2799032A1 (en) 2013-04-29 2014-11-05 3M Innovative Properties Company A method of capturing data from a patient's dentition and system for performing such method
US8897526B2 (en) 2011-05-06 2014-11-25 Sirona Dental Systems Gmbh Method, system, and computer-readable medium for uncovering and planning an accurate dental preparation
US8903746B2 (en) 2012-03-22 2014-12-02 Audrey Kudritskiy System and method for viewing, modifying, storing, and running artificial neural network components
US8903476B2 (en) 2009-03-08 2014-12-02 Oprobe, Llc Multi-function optical probe system for medical and veterinary applications
US9084568B2 (en) 2009-08-05 2015-07-21 Telesystems Co., Ltd. Radiation imaging apparatus and imaging method using radiation
US9107723B2 (en) 2005-11-22 2015-08-18 Benson Luther Hall System and method for the design, creation and installation of implant-supported dental prostheses
US9173727B2 (en) 2007-07-06 2015-11-03 Shofu Inc. Shade guide, method for discriminating tooth colors, artificial tooth manufacturing method
US9185388B2 (en) 2010-11-03 2015-11-10 3Dmedia Corporation Methods, systems, and computer program products for creating three-dimensional video sequences
US9208612B2 (en) 2010-02-12 2015-12-08 The University Of North Carolina At Chapel Hill Systems and methods that generate height map models for efficient three dimensional reconstruction from depth information
US9262864B2 (en) 2007-06-29 2016-02-16 3M Innovative Properties Company Synchronized views of video data and three-dimensional model data
US9322646B2 (en) 2010-04-09 2016-04-26 The Trustees Of The Stevens Institute Of Technology Adaptive mechanism control and scanner positioning for improved three-dimensional laser scanning
US9402601B1 (en) 1999-06-22 2016-08-02 Teratech Corporation Methods for controlling an ultrasound imaging procedure and providing ultrasound images to an external non-ultrasound application via a network
US9456963B2 (en) 2012-12-18 2016-10-04 Dentca, Inc. Photo-curable resin compositions and method of using the same in three-dimensional printing for manufacturing artificial teeth and denture base
US9554857B2 (en) 2009-09-14 2017-01-31 Memorial Sloan-Kettering Cancer Center Apparatus, system and method for providing laser steering and focusing for incision, excision and ablation of tissue in minimally-invasive surgery
US9554692B2 (en) 2009-06-18 2017-01-31 EndoChoice Innovation Ctr. Ltd. Multi-camera endoscope
US9566138B2 (en) * 2010-10-01 2017-02-14 Rune Fisker Modeling and manufacturing of dentures
US9662188B2 (en) 2011-02-14 2017-05-30 Ivoclar Vivadent Ag Method for producing a dental restoration part and CAD/CAM device
US9675432B2 (en) * 2009-05-19 2017-06-13 Dentca, Inc. Method and apparatus for preparing removable dental prosthesis
US9707061B2 (en) * 2013-12-27 2017-07-18 James R. Glidewell Dental Ceramics, Inc. Apparatus and methods of making denture devices
US9827076B2 (en) 2011-12-22 2017-11-28 3M Innovative Properties Company Method and system for making a dental restoration
US9845745B2 (en) 2015-07-08 2017-12-19 Ford Global Technologies, Llc EVAP system with valve to improve canister purging
US9844430B2 (en) * 2013-12-27 2017-12-19 James R. Glidewell Dental Ceramics, Inc. Apparatus and methods of making denture devices
US9861457B2 (en) 2009-03-20 2018-01-09 3Shape A/S System and method for effective planning, visualization, and optimization of dental restorations
US20180153664A1 (en) 2014-02-07 2018-06-07 3Shape A/S Detecting tooth shade
US10111714B2 (en) 2014-01-27 2018-10-30 Align Technology, Inc. Adhesive objects for improving image registration of intraoral images
USRE48221E1 (en) 2010-12-06 2020-09-22 3Shape A/S System with 3D user interface integration
US10835361B2 (en) 2016-02-24 2020-11-17 3Shape A/S Detecting and monitoring development of a dental condition

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE10321863A1 (en) 2003-05-14 2004-12-30 Willy Voit Gmbh & Co. Stanz- Und Metallwerk Pipe, in particular to be used as side shaft of drive shaft, produced from flat metal sheet with seam closed by laser welding
US7773074B2 (en) 2005-06-28 2010-08-10 Siemens Medical Solutions Usa, Inc. Medical diagnostic imaging three dimensional navigation device and methods
JP2009116402A (en) 2007-11-01 2009-05-28 Canon Inc Print controller, print control method and print control program
DE102013020445B4 (en) * 2013-12-06 2016-02-18 Patio-K Ag Method and device for applying paint in the field of dental technology

Patent Citations (389)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3878905A (en) 1972-04-07 1975-04-22 Hawera Probst Kg Hartmetall Drill, especially rock drill
US3971065A (en) 1975-03-05 1976-07-20 Eastman Kodak Company Color imaging array
US4349880A (en) 1979-03-19 1982-09-14 Rca Corporation Inspection system for detecting defects in regular patterns
US4291958A (en) 1980-10-03 1981-09-29 Eastman Kodak Company Camera with electronic flash and piezoelectric lens motor
US4342227A (en) 1980-12-24 1982-08-03 International Business Machines Corporation Planar semiconductor three direction acceleration detecting device and method of fabrication
US4575805A (en) 1980-12-24 1986-03-11 Moermann Werner H Method and apparatus for the fabrication of custom-shaped implants
US4516231A (en) 1982-08-26 1985-05-07 Rca Corporation Optical disc system having momentum compensation
US4629324A (en) 1983-12-29 1986-12-16 Robotic Vision Systems, Inc. Arrangement for measuring depth based on lens focusing
US4640620A (en) 1983-12-29 1987-02-03 Robotic Vision Systems, Inc. Arrangement for rapid depth measurement using lens focusing
JPS62100716A (en) 1985-10-29 1987-05-11 Matsushita Electric Ind Co Ltd Photographing device
US4802759A (en) 1986-08-11 1989-02-07 Goro Matsumoto Three-dimensional shape measuring apparatus
US4781448A (en) 1987-03-02 1988-11-01 Medical Concepts Inc. Zoom lens adapter for endoscopic camera
WO1988007695A1 (en) 1987-03-27 1988-10-06 The Board Of Trustees Of The Leland Stanford Junio Scanning confocal optical microscope
US4896015A (en) 1988-07-29 1990-01-23 Refractive Laser Research & Development Program, Ltd. Laser delivery system
US5372502A (en) 1988-09-02 1994-12-13 Kaltenbach & Voight Gmbh & Co. Optical probe and method for the three-dimensional surveying of teeth
US5269325A (en) 1989-05-26 1993-12-14 Biomagnetic Technologies, Inc. Analysis of biological signals using data from arrays of sensors
US5151609A (en) 1989-08-02 1992-09-29 Hitachi, Ltd. Method of detecting solid shape of object with autofocusing and image detection at each focus level
US5181181A (en) 1990-09-27 1993-01-19 Triton Technologies, Inc. Computer apparatus input device for three-dimensional information
WO1992014118A1 (en) 1991-02-12 1992-08-20 Oxford Sensor Technology Limited An optical sensor
JPH06505096A (en) 1991-02-12 1994-06-09 オックスフォード・センサ・テクノロジイ・リミテッド light sensor
US5381236A (en) 1991-02-12 1995-01-10 Oxford Sensor Technology Limited Optical sensor for imaging an object
WO1992015034A1 (en) 1991-02-19 1992-09-03 Phoenix Laser Systems, Inc. Target movement compensation for laser surgical system
CN1067573A (en) 1991-02-19 1993-01-06 凤凰激光系统有限公司 The system and method for detection, correction and measurement target tissue
US5131844A (en) 1991-04-08 1992-07-21 Foster-Miller, Inc. Contact digitizer, particularly for dental applications
US6485413B1 (en) 1991-04-29 2002-11-26 The General Hospital Corporation Methods and apparatus for forward-directed optical scanning instruments
US5377011A (en) 1991-09-06 1994-12-27 Koch; Stephen K. Scanning system for three-dimensional object digitizing
US5339154A (en) 1991-10-15 1994-08-16 Kaltenbach & Voigt Gmb & Co. Method and apparatus for optical measurement of objects
US5428450A (en) 1992-12-22 1995-06-27 Bertin & Cie Method and apparatus for determining the color of an object that is transparent, diffusing, and absorbent, such as a tooth, in particular
JP3321866B2 (en) 1992-12-28 2002-09-09 株式会社日立製作所 Surface shape detecting apparatus and method
JPH06201337A (en) 1992-12-28 1994-07-19 Hitachi Ltd Surface shape detection device and its method
US5455899A (en) 1992-12-31 1995-10-03 International Business Machines Corporation High speed image data processing circuit
US5563343A (en) 1993-05-26 1996-10-08 Cornell Research Foundation, Inc. Microelectromechanical lateral accelerometer
US5737339A (en) 1994-02-08 1998-04-07 Fujitsu Limited Method and apparatus for verifying stored data
US5850289A (en) 1994-08-24 1998-12-15 Tricorder Technology Plc Scanning arrangement and method
US5615003A (en) 1994-11-29 1997-03-25 Hermary; Alexander T. Electromagnetic profile scanner
US5675407A (en) 1995-03-02 1997-10-07 Zheng Jason Geng Color ranging method for high speed low-cost three dimensional surface profile measurement
US5605459A (en) 1995-04-14 1997-02-25 Unisn Incorporated Method of and apparatus for making a dental set-up model
US5702249A (en) 1995-05-19 1997-12-30 Cooper; David H. Modular intra-oral imaging system video camera
US6229913B1 (en) 1995-06-07 2001-05-08 The Trustees Of Columbia University In The City Of New York Apparatus and methods for determining the three-dimensional shape of an object using active illumination and relative blurring in two-images due to defocus
US5766006A (en) 1995-06-26 1998-06-16 Murljacic; Maryann Lehmann Tooth shade analyzer system and methods
US7946845B2 (en) 1995-06-26 2011-05-24 Shade Analyzing Technologies, Inc. Tooth shade analyzer system and methods
WO1997002788A1 (en) 1995-07-07 1997-01-30 Siemens Aktiengesellschaft Process and device for computer-assisted restoration of teeth
DE19524855A1 (en) 1995-07-07 1997-01-09 Siemens Ag Method and device for computer-aided restoration of teeth
EP0837659A1 (en) 1995-07-07 1998-04-29 Siemens Aktiengesellschaft Process and device for computer-assisted restoration of teeth
US5737084A (en) 1995-09-29 1998-04-07 Takaoka Electric Mtg. Co., Ltd. Three-dimensional shape measuring apparatus
WO1997014932A1 (en) 1995-10-20 1997-04-24 Optronic Consult Ab Process and device for the measuring of a three-dimensional shape
US5851113A (en) 1996-01-02 1998-12-22 Lj Laboratories, L.L.C. Apparatus and method for measuring the color of teeth
US7215430B2 (en) 1996-04-24 2007-05-08 Leica Geosystems Hds Llc Integrated system for quickly and accurately imaging and modeling three-dimensional objects
US6135961A (en) 1996-06-28 2000-10-24 Sonosite, Inc. Ultrasonic signal processor for a hand held ultrasonic diagnostic instrument
US5722412A (en) 1996-06-28 1998-03-03 Advanced Technology Laboratories, Inc. Hand held ultrasonic diagnostic instrument
US20050090749A1 (en) 1996-09-02 2005-04-28 Rudger Rubbert Method and device for carrying out optical pick up
US6007332A (en) 1996-09-26 1999-12-28 O'brien; William J. Tooth color matching system
US6263234B1 (en) 1996-10-01 2001-07-17 Leica Microsystems Heidelberg Gmbh Confocal surface-measuring device
DE19642247C1 (en) 1996-10-12 1998-01-15 Sebastian Meller Prosthesis production system for prepared tooth
US6334773B1 (en) 1997-02-24 2002-01-01 Dentronic Ab Method and arrangement for making artificial teeth
WO1998045745A1 (en) 1997-04-04 1998-10-15 Isis Innovation Limited Microscopy imaging apparatus and method
US6259452B1 (en) 1997-04-14 2001-07-10 Massachusetts Institute Of Technology Image drawing system and method with real-time occlusion culling
US6334853B1 (en) 1997-05-22 2002-01-01 Cadent Ltd Method for obtaining a dental occlusion map
US6249616B1 (en) 1997-05-30 2001-06-19 Enroute, Inc Combining digital images based on three-dimensional relationships between source image data sets
US6450807B1 (en) 1997-06-20 2002-09-17 Align Technology, Inc. System and method for positioning teeth
US7134874B2 (en) 1997-06-20 2006-11-14 Align Technology, Inc. Computer automated development of an orthodontic treatment plan and appliance
US6471511B1 (en) 1997-06-20 2002-10-29 Align Technology, Inc. Defining tooth-moving appliances computationally
US6148120A (en) 1997-10-30 2000-11-14 Cognex Corporation Warping of focal images to correct correspondence error
US6026189A (en) 1997-11-13 2000-02-15 National Research Council Of Canada Method of recognizing objects within two-dimensional and three-dimensional images
WO1999047964A1 (en) 1998-03-20 1999-09-23 Genetic Microsystems, Inc. Wide field of view and high speed scanning microscopy
US6206691B1 (en) 1998-05-20 2001-03-27 Shade Analyzing Technologies, Inc. System and methods for analyzing tooth shades
US6081739A (en) 1998-05-21 2000-06-27 Lemchen; Marc S. Scanning device or methodology to produce an image incorporating correlated superficial, three dimensional surface and x-ray images and measurements of an object
US6072496A (en) 1998-06-08 2000-06-06 Microsoft Corporation Method and system for capturing and representing 3D geometry, color and shading of facial expressions and other animated objects
US20010030748A1 (en) 1998-07-09 2001-10-18 Jung Wayne D. Apparatus and method for measuring optical characteristics of an object
WO2000008415A1 (en) 1998-08-05 2000-02-17 Cadent Ltd. Imaging a three-dimensional structure by confocal focussing an array of light beams
US20070109559A1 (en) 1998-08-05 2007-05-17 Cadent Ltd Method and apparatus for imaging three-dimensional structure
US20060158665A1 (en) 1998-08-05 2006-07-20 Cadent Ltd. Method and apparatus for imaging three-dimensional structure
US6697164B1 (en) 1998-08-05 2004-02-24 Cadent Ltd. Imaging a three-dimensional structure by confocal focussing an array of light beams
US6967644B1 (en) 1998-10-01 2005-11-22 Canon Kabushiki Kaisha Coordinate input apparatus and control method thereof, and computer readable memory
US20070026363A1 (en) 1998-11-03 2007-02-01 Maryann Lehmann Interactive dental restorative network
US6575751B1 (en) 1998-11-03 2003-06-10 Shade Analyzing Technologies, Inc. Interactive dental restorative network
US6361489B1 (en) 1998-11-25 2002-03-26 Jory Tsai Medical inspection device
US7068825B2 (en) 1999-03-08 2006-06-27 Orametrix, Inc. Scanning system and calibration method for capturing precise three-dimensional information of objects
US7058213B2 (en) 1999-03-08 2006-06-06 Orametrix, Inc. Scanning system and calibration method for capturing precise three-dimensional information of objects
US7099732B2 (en) 1999-03-29 2006-08-29 Genex Technologies, Inc. Sanitary sleeve or tip for intra-oral three-dimensional camera
US6227850B1 (en) 1999-05-13 2001-05-08 Align Technology, Inc. Teeth viewing system
US6751344B1 (en) 1999-05-28 2004-06-15 Champion Orthotic Investments, Inc. Enhanced projector system for machine vision
US9402601B1 (en) 1999-06-22 2016-08-02 Teratech Corporation Methods for controlling an ultrasound imaging procedure and providing ultrasound images to an external non-ultrasound application via a network
WO2001011193A1 (en) 1999-08-05 2001-02-15 Jama Mining Equipment Ab Arrangement for fixing rock bolts during rock reinforcement
US6251073B1 (en) 1999-08-20 2001-06-26 Novasonics, Inc. Miniaturized ultrasound apparatus and method
US20030096210A1 (en) 1999-11-30 2003-05-22 Orametrix, Inc. Interactive orthodontic care system based on intra-oral scanning of teeth
US7296996B2 (en) 1999-11-30 2007-11-20 Orametrix, Inc. Virtual bracket placement and evaluation
US8121718B2 (en) 1999-11-30 2012-02-21 Orametrix, Inc. Interactive orthodontic care system based on intra-oral scanning of teeth
US8998608B2 (en) * 1999-11-30 2015-04-07 Orametrix, Inc. Three-dimensional occlusal and interproximal contact detection and display using virtual tooth models
US7160110B2 (en) 1999-11-30 2007-01-09 Orametrix, Inc. Three-dimensional occlusal and interproximal contact detection and display using virtual tooth models
US7458812B2 (en) * 1999-11-30 2008-12-02 Orametrix, Inc. Three-dimensional occlusal and interproximal contact detection and display using virtual tooth models
US20090291417A1 (en) 1999-11-30 2009-11-26 Rubbert Ruedger Interactive orthodontic care system based on intra-oral scanning of teeth
US6990228B1 (en) 1999-12-17 2006-01-24 Canon Kabushiki Kaisha Image processing apparatus
US6476803B1 (en) 2000-01-06 2002-11-05 Microsoft Corporation Object modeling system and process employing noise elimination and robust surface extraction techniques
US6865289B1 (en) 2000-02-07 2005-03-08 Canon Kabushiki Kaisha Detection and removal of image occlusion errors
US20040125103A1 (en) 2000-02-25 2004-07-01 Kaufman Arie E. Apparatus and method for volume processing and rendering
US7197179B2 (en) 2000-04-28 2007-03-27 Orametrix, Inc. Methods for registration of three-dimensional frames to create three-dimensional virtual models of objects
US6532299B1 (en) 2000-04-28 2003-03-11 Orametrix, Inc. System and method for mapping a surface
US7471821B2 (en) 2000-04-28 2008-12-30 Orametrix, Inc. Method and apparatus for registering a known digital object to scanned 3-D model
WO2001084479A1 (en) 2000-04-28 2001-11-08 Orametirix, Inc. Method and system for scanning a surface and generating a three-dimensional object
US7027642B2 (en) 2000-04-28 2006-04-11 Orametrix, Inc. Methods for registration of three-dimensional frames to create three-dimensional virtual models of objects
US20070081718A1 (en) 2000-04-28 2007-04-12 Rudger Rubbert Methods for registration of three-dimensional frames to create three-dimensional virtual models of objects
US6743014B2 (en) 2000-05-16 2004-06-01 Ivoclar Vivadent Ag Shade determination apparatus and method for specifying and determining colors for teeth and dental restorations
US6975898B2 (en) 2000-06-19 2005-12-13 University Of Washington Medical imaging, diagnosis, and therapy using a scanning single optical fiber system
US6750873B1 (en) 2000-06-27 2004-06-15 International Business Machines Corporation High quality texture reconstruction from multiple scans
US20030164952A1 (en) 2000-08-25 2003-09-04 Nikolaj Deichmann Method and apparatus for three-dimensional optical scanning of interior surfaces
US7079679B2 (en) 2000-09-27 2006-07-18 Canon Kabushiki Kaisha Image processing apparatus
US6592371B2 (en) 2000-10-25 2003-07-15 Duane Durbin Method and system for imaging and modeling a three dimensional structure
US6954550B2 (en) 2000-11-29 2005-10-11 Omron Corporation Image processing method and apparatus
US6645148B2 (en) 2001-03-20 2003-11-11 Vermon Ultrasonic probe including pointing devices for remotely controlling functions of an associated imaging system
WO2002076327A1 (en) 2001-03-23 2002-10-03 Decim Ab A method of and an arrangement for a dental restoration
US20120015316A1 (en) 2001-04-13 2012-01-19 Rohit Sachdeva Unified three dimensional virtual craniofacial and dentition model and uses thereof
US8177551B2 (en) 2001-04-13 2012-05-15 Orametrix, Inc. Method and system for comprehensive evaluation of orthodontic treatment using unified workstation
US8469705B2 (en) 2001-04-13 2013-06-25 Orametrix, Inc. Method and system for integrated orthodontic treatment planning using unified workstation
US20060025684A1 (en) 2001-04-19 2006-02-02 Sonosite, Inc. Medical diagnostic ultrasound instrument with ECG module, authorization mechanism and methods of use
US7086863B2 (en) 2001-04-23 2006-08-08 Cicero Dental Systems, B.V. Method for production of an artificial tooth
US8078006B1 (en) 2001-05-04 2011-12-13 Legend3D, Inc. Minimal artifact image sequence depth enhancement system and method
US7010223B2 (en) 2001-05-26 2006-03-07 Dürr Dental GmbH & Co. KG Dental or endoscopic camera
US7213214B2 (en) 2001-06-12 2007-05-01 Idelix Software Inc. Graphical user interface with zoom for detail-in-context presentations
US20030043089A1 (en) 2001-08-17 2003-03-06 Eric Hanson Doubling of speed in CMOS sensor with column-parallel ADCS
US6904159B2 (en) 2001-12-20 2005-06-07 Mitsubishi Electric Research Laboratories, Inc. Identifying moving objects in a video using volume growing and change detection masks
WO2003060587A1 (en) 2002-01-15 2003-07-24 École Polytechnique Fédérale de Lausanne Microscopy imaging apparatus and mehod for generating an image
US7141020B2 (en) 2002-02-20 2006-11-28 Koninklijke Philips Electronics N.V. Portable 3D ultrasound system
US20030158482A1 (en) 2002-02-20 2003-08-21 Koninklijke Philips Electronics N.V. Portable 3D ultrasound system
WO2003073457A2 (en) 2002-02-21 2003-09-04 Lj Laboratories Llc Miniaturized system and method for measuring optical characteristics
US20030156283A1 (en) 2002-02-21 2003-08-21 Lj Laboratories, L.L.C. Miniaturized system and method for measuring optical characteristics
US7831292B2 (en) 2002-03-06 2010-11-09 Mako Surgical Corp. Guidance system and method for surgical procedures with improved feedback
US7166537B2 (en) 2002-03-18 2007-01-23 Sarcos Investments Lc Miniaturized imaging device with integrated circuit connector system
US7636455B2 (en) 2002-06-04 2009-12-22 Raytheon Company Digital image edge detection and road network tracking method and system
US6761561B2 (en) 2002-06-07 2004-07-13 Schick Technologies Wireless dental camera
US7385708B2 (en) 2002-06-07 2008-06-10 The University Of North Carolina At Chapel Hill Methods and systems for laser based real-time structured light depth extraction
JP2004029685A (en) 2002-06-23 2004-01-29 Akira Ishii Method and device for fixed magnification imaging using variable focal lens
US7077647B2 (en) 2002-08-22 2006-07-18 Align Technology, Inc. Systems and methods for treatment analysis by teeth matching
US20040155975A1 (en) 2002-09-17 2004-08-12 Hart Douglas P. 3-D imaging system
US20070041729A1 (en) 2002-10-23 2007-02-22 Philip Heinz Systems and methods for detecting changes in incident optical radiation at high frequencies
US7230771B2 (en) 2002-10-25 2007-06-12 Koninklijke Philips Electronics N.V. Zoom lens
US20040080754A1 (en) 2002-10-29 2004-04-29 Mitutoyo Corporation Interferometer using integrated imaging array and high-density phase-shifting array
US7123760B2 (en) 2002-11-21 2006-10-17 General Electric Company Method and apparatus for removing obstructing structures in CT imaging
US6977732B2 (en) 2002-12-26 2005-12-20 National Taiwan University Miniature three-dimensional contour scanner
WO2004066615A1 (en) 2003-01-22 2004-08-05 Nokia Corporation Image control
US20060146009A1 (en) 2003-01-22 2006-07-06 Hanno Syrbe Image control
US20060072123A1 (en) 2003-01-25 2006-04-06 Wilson John E Methods and apparatus for making images including depth information
US20040185422A1 (en) 2003-03-21 2004-09-23 Sirona Dental Systems Gmbh Data base, tooth model and restorative item constructed from digitized images of real teeth
US20040254476A1 (en) 2003-03-24 2004-12-16 Henley Quadling Laser digitizer system for dental applications
US7184150B2 (en) 2003-03-24 2007-02-27 D4D Technologies, Llc Laser digitizer system for dental applications
US20040204787A1 (en) 2003-04-03 2004-10-14 Avi Kopelman Method and system for fabricating a dental coping, and a coping fabricated thereby
US20050020910A1 (en) 2003-04-30 2005-01-27 Henley Quadling Intra-oral imaging system
US7355721B2 (en) 2003-05-05 2008-04-08 D4D Technologies, Llc Optical coherence tomography imaging
DE10321883A1 (en) 2003-05-07 2004-12-09 Universität Stuttgart Triangulation measurement device for determining object 3D structure has illumination and observation arrays with a projected pattern being evaluated using cross correlation or phase shift analysis
US7339170B2 (en) 2003-07-16 2008-03-04 Shrenik Deliwala Optical encoding and reconstruction
US20050057745A1 (en) 2003-09-17 2005-03-17 Bontje Douglas A. Measurement methods and apparatus
JP2005098833A (en) 2003-09-25 2005-04-14 Keyence Corp Displacement meter and displacement measuring method
US20050074718A1 (en) 2003-10-03 2005-04-07 Pye Graham Tooth shade scan system and method
US7349104B2 (en) 2003-10-23 2008-03-25 Technest Holdings, Inc. System and a method for three-dimensional imaging systems
US20080058783A1 (en) 2003-11-04 2008-03-06 Palomar Medical Technologies, Inc. Handheld Photocosmetic Device
US7221332B2 (en) 2003-12-19 2007-05-22 Eastman Kodak Company 3D stereo OLED display
US20050142517A1 (en) 2003-12-30 2005-06-30 Howard Frysh System for producing a dental implant and method
WO2005067389A2 (en) 2004-01-15 2005-07-28 Technion Research & Development Foundation Ltd. Three-dimensional video scanner
US20060251408A1 (en) 2004-01-23 2006-11-09 Olympus Corporation Image processing system and camera
US20100201986A1 (en) 2004-02-20 2010-08-12 Jean-Marc Inglese Equipment and method for measuring dental shade
US8400635B2 (en) 2004-02-20 2013-03-19 Carestream Health, Inc. Equipment and method for measuring dental shade
US20070194214A1 (en) 2004-03-19 2007-08-23 Sirona Dental Systems Gmbh High-Speed Measuring Device And Method Based On A Confocal Microscopy Principle
CN1934481A (en) 2004-03-19 2007-03-21 塞隆纳牙科系统有限责任公司 High-speed measuring device and method based on a confocal microscopy principle
US20050212753A1 (en) 2004-03-23 2005-09-29 Marvit David L Motion controlled remote controller
US20050212756A1 (en) 2004-03-23 2005-09-29 Marvit David L Gesture based navigation of a handheld user interface
US20080132886A1 (en) 2004-04-09 2008-06-05 Palomar Medical Technologies, Inc. Use of fractional emr technology on incisions and internal tissues
US20050232509A1 (en) 2004-04-16 2005-10-20 Andrew Blake Virtual image artifact detection
CN1906678A (en) 2004-04-21 2007-01-31 松下电器产业株式会社 Confocal optical system aperture position detector, confocal optical system aperture position controller, optical head, optical information processor and confocal optical system aperture position dete
US20050237581A1 (en) 2004-04-21 2005-10-27 Knighton Mark S Hand held portable three dimensional scanner
US20080316898A1 (en) 2004-04-21 2008-12-25 Matsushita Electric Industrial Co., Ltd. Confocal Optical System Aperture Position Detector, Confocal Optical System Aperature Position Controller, Optical Head, Optical Information Processor, and Confocal Optical System Aperture Position Detecting Method
US20050243330A1 (en) 2004-04-28 2005-11-03 Simon Magarill Methods and apparatus for determining three dimensional configurations
US7483062B2 (en) 2004-04-28 2009-01-27 International Business Machines Corporation Method for removal of moving objects from a video stream
US20070182812A1 (en) 2004-05-19 2007-08-09 Ritchey Kurtis J Panoramic image-based virtual reality/telepresence audio-visual system and method
US20050283065A1 (en) 2004-06-17 2005-12-22 Noam Babayoff Method for providing data associated with the intraoral cavity
US8675207B2 (en) * 2004-06-17 2014-03-18 Cadent Ltd. Method and apparatus for colour imaging a three-dimensional structure
US7319529B2 (en) * 2004-06-17 2008-01-15 Cadent Ltd Method and apparatus for colour imaging a three-dimensional structure
US9404740B2 (en) * 2004-06-17 2016-08-02 Align Technology, Inc. Method and apparatus for colour imaging a three-dimensional structure
US9101433B2 (en) * 2004-06-17 2015-08-11 Align Technology, Inc. Method and apparatus for colour imaging a three-dimensional structure
US8885175B2 (en) * 2004-06-17 2014-11-11 Cadent Ltd. Method and apparatus for colour imaging a three-dimensional structure
US8363228B2 (en) 2004-06-17 2013-01-29 Cadent Ltd. Method and apparatus for colour imaging a three-dimensional structure
US20080024768A1 (en) 2004-06-17 2008-01-31 Cadent Ltd Method and apparatus for colour imaging a three-dimensional structure
US8451456B2 (en) 2004-06-17 2013-05-28 Cadent Ltd. Method and apparatus for colour imaging a three-dimensional structure
US20060001739A1 (en) 2004-06-17 2006-01-05 Noam Babayoff Method and apparatus for colour imaging a three-dimensional structure
US7698068B2 (en) 2004-06-17 2010-04-13 Cadent Ltd. Method for providing data associated with the intraoral cavity
US7724378B2 (en) * 2004-06-17 2010-05-25 Cadent Ltd. Method and apparatus for colour imaging a three-dimensional structure
US20060020204A1 (en) 2004-07-01 2006-01-26 Bracco Imaging, S.P.A. System and method for three-dimensional space management and visualization of ultrasound data ("SonoDEX")
US8331653B2 (en) 2004-08-11 2012-12-11 Tokyo Institute Of Technology Object detector
US8180100B2 (en) 2004-08-11 2012-05-15 Honda Motor Co., Ltd. Plane detector and detecting method
US7522322B2 (en) 2004-08-19 2009-04-21 Carestream Health, Inc. Apparatus for dental shade measurement
US7762814B2 (en) 2004-09-14 2010-07-27 Oratio B.V. Method of manufacturing and installing a ceramic dental implant with an aesthetic implant abutment
US20070252074A1 (en) 2004-10-01 2007-11-01 The Board Of Trustees Of The Leland Stanford Junio Imaging Arrangements and Methods Therefor
CN101426085A (en) 2004-10-01 2009-05-06 科兰·斯坦福青年大学托管委员会 Imaging arrangements and methods therefor
US20060072189A1 (en) 2004-10-06 2006-04-06 Dimarzio Charles A Confocal reflectance microscope system with dual rotating wedge scanner assembly
US20060092133A1 (en) 2004-11-02 2006-05-04 Pierre A. Touma 3D mouse and game controller based on spherical coordinates system and system for use
US8026916B2 (en) 2004-12-14 2011-09-27 Align Technology, Inc. Image-based viewing system
US20060127852A1 (en) 2004-12-14 2006-06-15 Huafeng Wen Image based orthodontic treatment viewing system
WO2006065955A2 (en) 2004-12-14 2006-06-22 Orthoclear Holdings, Inc. Image based orthodontic treatment methods
US20090040175A1 (en) 2004-12-22 2009-02-12 Rex Fang Xu Input interface device with transformable form factor
US7494338B2 (en) 2005-01-11 2009-02-24 Duane Durbin 3D dental scanner
US20090167948A1 (en) 2005-02-08 2009-07-02 Berman Steven T System and Method for Selective Image Capture, Transmission and Reconstruction
US20060212260A1 (en) 2005-03-03 2006-09-21 Cadent Ltd. System and method for scanning an intraoral cavity
US20090097108A1 (en) 2005-05-12 2009-04-16 Fox William J Confocal scanning microscope having optical and scanning systems which provide a handheld imaging head
US7609875B2 (en) 2005-05-27 2009-10-27 Orametrix, Inc. Scanner system and method for mapping surface of three-dimensional object
US7474414B2 (en) 2005-06-01 2009-01-06 Inus Technology, Inc. System and method of guiding real-time inspection using 3D scanners
US20070015025A1 (en) 2005-07-15 2007-01-18 Honda Motor Co., Ltd. Membrane electrode assembly for solid polymer electrolyte fuel cell
US7551353B2 (en) 2005-07-27 2009-06-23 Samsung Electronics Co., Ltd. Glassless stereoscopic display
US20070031774A1 (en) 2005-08-03 2007-02-08 3M Innovative Properties Company Registering physical and virtual tooth structures with markers
US20070064242A1 (en) 2005-09-01 2007-03-22 Sonaray, Inc. Method and system for obtaining high resolution 3-D images of moving objects by use of sensor fusion
JP2007072103A (en) 2005-09-06 2007-03-22 Fujifilm Holdings Corp Camera
US8625854B2 (en) 2005-09-09 2014-01-07 Industrial Research Limited 3D scene scanner and a position and orientation system
US20070078340A1 (en) 2005-09-30 2007-04-05 Siemens Medical Solutions Usa, Inc. Method and apparatus for controlling ultrasound imaging systems having positionable transducers
US8390821B2 (en) 2005-10-11 2013-03-05 Primesense Ltd. Three-dimensional sensing using speckle patterns
US7605817B2 (en) 2005-11-09 2009-10-20 3M Innovative Properties Company Determining camera motion
US7929751B2 (en) 2005-11-09 2011-04-19 Gi, Llc Method and apparatus for absolute-coordinate three-dimensional surface imaging
US20070103460A1 (en) 2005-11-09 2007-05-10 Tong Zhang Determining camera motion
US9107723B2 (en) 2005-11-22 2015-08-18 Benson Luther Hall System and method for the design, creation and installation of implant-supported dental prostheses
US20070134615A1 (en) 2005-12-08 2007-06-14 Lovely Peter S Infrared dental imaging
US20070140539A1 (en) 2005-12-19 2007-06-21 Olympus Corporation Image combining apparatus
US20090076321A1 (en) 2005-12-26 2009-03-19 Kanagawa Furniture Co., Ltd. Digital camera for taking image inside oral cavity
US9208531B2 (en) 2006-01-20 2015-12-08 3M Innovative Properties Company Digital dentistry
US7813591B2 (en) 2006-01-20 2010-10-12 3M Innovative Properties Company Visual feedback of 3D scan parameters
US20070171220A1 (en) 2006-01-20 2007-07-26 Kriveshko Ilya A Three-dimensional scan recovery
US7840042B2 (en) 2006-01-20 2010-11-23 3M Innovative Properties Company Superposition for visualization of three-dimensional data acquisition
US20090298017A1 (en) 2006-01-20 2009-12-03 3M Innivative Properties Company Digital dentistry
US7940260B2 (en) 2006-01-20 2011-05-10 3M Innovative Properties Company Three-dimensional scan recovery
WO2007084727A1 (en) 2006-01-20 2007-07-26 3M Innovative Properties Company Digital dentistry
US8035637B2 (en) 2006-01-20 2011-10-11 3M Innovative Properties Company Three-dimensional scan recovery
US20070172112A1 (en) 2006-01-20 2007-07-26 Paley Eric B Visual feedback of 3d scan parameters
US20070212667A1 (en) 2006-03-13 2007-09-13 Jung Wayne D Systems and methods for preparing dental restorations
US20100009308A1 (en) 2006-05-05 2010-01-14 Align Technology, Inc. Visualizing and Manipulating Digital Models for Dental Treatment
US7460248B2 (en) 2006-05-15 2008-12-02 Carestream Health, Inc. Tissue imaging system
US8384665B1 (en) 2006-07-14 2013-02-26 Ailive, Inc. Method and system for making a selection in 3D virtual environment
US20090177050A1 (en) 2006-07-17 2009-07-09 Medrad, Inc. Integrated medical imaging systems
US8442283B2 (en) 2006-08-30 2013-05-14 Anatomage Inc. Patient-specific three-dimensional dentition model
US20080063998A1 (en) 2006-09-12 2008-03-13 Rongguang Liang Apparatus for caries detection
US20080070684A1 (en) 2006-09-14 2008-03-20 Mark Haigh-Hutchinson Method and apparatus for using a common pointing input to control 3D viewpoint and object targeting
US20080071143A1 (en) 2006-09-18 2008-03-20 Abhishek Gattani Multi-dimensional navigation of endoscopic video
US7708557B2 (en) 2006-10-16 2010-05-04 Natural Dental Implants Ag Customized dental prosthesis for periodontal- or osseointegration, and related systems and methods
US8090194B2 (en) 2006-11-21 2012-01-03 Mantis Vision Ltd. 3D geometric modeling and motion capture using both single and dual imaging
US20080118886A1 (en) 2006-11-21 2008-05-22 Rongguang Liang Apparatus for dental oct imaging
US20080131028A1 (en) 2006-11-30 2008-06-05 Pillman Bruce H Producing low resolution images
US20080194928A1 (en) 2007-01-05 2008-08-14 Jadran Bandic System, device, and method for dermal imaging
DE102007005726A1 (en) 2007-01-31 2008-08-07 Sirona Dental Systems Gmbh Device and method for 3D optical measurement
US20090279103A1 (en) 2007-01-31 2009-11-12 Sirona Dental Systems Gmbh Apparatus and method for optical 3d measurement
JP2008194108A (en) 2007-02-09 2008-08-28 Shiyoufuu:Kk Three-dimensional characteristic measuring and displaying apparatus with positional direction detecting function
US20080194950A1 (en) 2007-02-13 2008-08-14 General Electric Company Ultrasound imaging remote control unit
US7550707B2 (en) 2007-03-01 2009-06-23 Sony Corporation Biometrics authentication system with authentication based upon multiple layers
US8335353B2 (en) 2007-04-04 2012-12-18 Sony Corporation Biometrics authentication system
WO2008125605A2 (en) 2007-04-13 2008-10-23 Michael Schwertner Method and assembly for optical reproduction with depth discrimination
US20100108873A1 (en) 2007-04-13 2010-05-06 Michael Schwertner Method and assembly for optical reproduction with depth discrimination
US9262864B2 (en) 2007-06-29 2016-02-16 3M Innovative Properties Company Synchronized views of video data and three-dimensional model data
US9173727B2 (en) 2007-07-06 2015-11-03 Shofu Inc. Shade guide, method for discriminating tooth colors, artificial tooth manufacturing method
US8003889B2 (en) 2007-08-02 2011-08-23 Thomas & Betts International, Inc. Conduit sleeve pass through for concrete construction
US20090087050A1 (en) 2007-08-16 2009-04-02 Michael Gandyra Device for determining the 3D coordinates of an object, in particular of a tooth
WO2009026645A1 (en) 2007-08-31 2009-03-05 Signostics Pty Ltd Apparatus and method for medical scanning
US20090061381A1 (en) 2007-09-05 2009-03-05 Duane Milford Durbin Systems and methods for 3D previewing
US20090322676A1 (en) 2007-09-07 2009-12-31 Apple Inc. Gui applications for use with 3d remote controller
WO2009034157A1 (en) 2007-09-12 2009-03-19 Degudent Gmbh Method for determining the position of an intraoral measuring device
US8456636B2 (en) 2007-10-01 2013-06-04 Ferton Holding, S.A. Device for detecting signs of bacterial infection of the teeth
US20090103103A1 (en) 2007-10-18 2009-04-23 Mht Optic Research Ag Device for tomographic scanning objects
JP2009098146A (en) 2007-10-18 2009-05-07 Mht Optic Research Ag System for tomographic scanning objects
US8144954B2 (en) 2007-11-08 2012-03-27 D4D Technologies, Llc Lighting compensated dynamic texture mapping of 3-D models
US8532355B2 (en) 2007-11-08 2013-09-10 D4D Technologies, Llc Lighting compensated dynamic texture mapping of 3-D models
US8280152B2 (en) 2007-11-15 2012-10-02 Sirona Dental Systems Gmbh Method for optical measurement of the three dimensional geometry of objects
WO2009063088A2 (en) 2007-11-15 2009-05-22 Sirona Dental Systems Gmbh Method for optical measurement of objects using a triangulation method
US20090133260A1 (en) 2007-11-26 2009-05-28 Ios Technologies, Inc 3D dental shade matching and apparatus
US20090160858A1 (en) 2007-12-21 2009-06-25 Industrial Technology Research Institute Method for reconstructing three dimensional model
US20090233253A1 (en) 2007-12-21 2009-09-17 Mrazek William R Dental shade guide
WO2009089126A1 (en) 2008-01-04 2009-07-16 3M Innovative Properties Company Three-dimensional model refinement
US7929151B2 (en) 2008-01-11 2011-04-19 Carestream Health, Inc. Intra-oral camera for diagnostic and cosmetic imaging
US8103134B2 (en) 2008-02-20 2012-01-24 Samsung Electronics Co., Ltd. Method and a handheld device for capturing motion
US20090217207A1 (en) 2008-02-22 2009-08-27 Robert Kagermeier Display of a medical image
CN101513350A (en) 2008-02-22 2009-08-26 西门子公司 Device and method for displaying medical image and imaging system
US8121351B2 (en) 2008-03-09 2012-02-21 Microsoft International Holdings B.V. Identification of objects in a 3D video using non/over reflective clothing
US20090231649A1 (en) 2008-03-12 2009-09-17 Sirat Gabriel Y Intraoral Imaging System and Method based on Conoscopic Holography
US8092215B2 (en) 2008-05-23 2012-01-10 Align Technology, Inc. Smile designer
US8386061B2 (en) 2008-06-02 2013-02-26 Dentsply International Inc. Methods for designing a customized dental prosthesis using digital images of a patient
US20110188726A1 (en) 2008-06-18 2011-08-04 Ram Nathaniel Method and system for stitching multiple images into a panoramic image
US8345961B2 (en) 2008-09-10 2013-01-01 Huawei Device Co., Ltd. Image stitching method and apparatus
US20100239136A1 (en) 2008-09-18 2010-09-23 Steinbichler Optotechnik Gmbh Device for Determining the 3D Coordinates of an Object, In Particular of a Tooth
US8743114B2 (en) 2008-09-22 2014-06-03 Intel Corporation Methods and systems to determine conservative view cell occlusion
US20100079581A1 (en) 2008-09-30 2010-04-01 Texas Instruments Incorporated 3d camera using flash with structured light
US20100085636A1 (en) 2008-10-06 2010-04-08 Mht Optic Research Ag Optical System for a Confocal Microscope
US8571397B2 (en) 2008-11-21 2013-10-29 Carestream Health, Inc. Auto focus intraoral camera with liquid lens
WO2010064156A1 (en) 2008-12-03 2010-06-10 Koninklijke Philips Electronics, N.V. Ultrasound assembly and system comprising interchangable transducers and displays
US20100157086A1 (en) 2008-12-15 2010-06-24 Illumina, Inc Dynamic autofocus method and system for assay imager
EP2200332A1 (en) 2008-12-17 2010-06-23 Robert Bosch GmbH Autostereoscopic display
US20100156901A1 (en) 2008-12-22 2010-06-24 Electronics And Telecommunications Research Institute Method and apparatus for reconstructing 3d model
US20110316978A1 (en) 2009-02-25 2011-12-29 Dimensional Photonics International, Inc. Intensity and color display for a three-dimensional metrology system
US8903476B2 (en) 2009-03-08 2014-12-02 Oprobe, Llc Multi-function optical probe system for medical and veterinary applications
US20100231509A1 (en) 2009-03-12 2010-09-16 Marc Boillot Sterile Networked Interface for Medical Systems
US9861457B2 (en) 2009-03-20 2018-01-09 3Shape A/S System and method for effective planning, visualization, and optimization of dental restorations
US8914245B2 (en) 2009-03-20 2014-12-16 Andrew David Hopkins Ultrasound probe with accelerometer
WO2010106379A1 (en) 2009-03-20 2010-09-23 Mediwatch Uk Limited Ultrasound probe with accelerometer
US8828287B2 (en) 2009-03-25 2014-09-09 Oratio B.V. Veneered dental restoration with a controlled shade
US20100268069A1 (en) 2009-04-16 2010-10-21 Rongguang Liang Dental surface imaging using polarized fringe projection
US8477320B2 (en) 2009-05-15 2013-07-02 Degudent Gmbh Method and measuring arrangement for the three-dimensional measurement of an object
US9675432B2 (en) * 2009-05-19 2017-06-13 Dentca, Inc. Method and apparatus for preparing removable dental prosthesis
US8564657B2 (en) 2009-05-29 2013-10-22 Honda Research Institute Europe Gmbh Object motion detection system based on combining 3D warping techniques and a proper object motion detection
US8570530B2 (en) 2009-06-03 2013-10-29 Carestream Health, Inc. Apparatus for dental surface shape and shade imaging
DE102009023952A1 (en) 2009-06-04 2010-12-09 DüRR DENTAL AG Method for determining tooth color of e.g. dental prosthesis, involves providing difference between color values of images and tooth color pattern as measurement for correlating colors of tooth region and tooth color pattern
US8878905B2 (en) 2009-06-17 2014-11-04 3Shape A/S Focus scanning apparatus
WO2010145669A1 (en) 2009-06-17 2010-12-23 3Shape A/S Focus scanning apparatus
US10349041B2 (en) 2009-06-17 2019-07-09 3Shape A/S Focus scanning apparatus
US10326982B2 (en) 2009-06-17 2019-06-18 3Shape A/S Focus scanning apparatus
US10349042B1 (en) 2009-06-17 2019-07-09 3Shape A/S Focus scanning apparatus
US20190200006A1 (en) 2009-06-17 2019-06-27 3Shape A/S Focus scanning apparatus
US20150054922A1 (en) 2009-06-17 2015-02-26 3Shape A/S Focus scanning apparatus
US10097815B2 (en) 2009-06-17 2018-10-09 3Shape A/S Focus scanning apparatus
US20190289283A1 (en) 2009-06-17 2019-09-19 3Shape A/S Focus scanning apparatus
US20180255293A1 (en) 2009-06-17 2018-09-06 3Shape A/S Focus scanning apparatus
US20220086418A1 (en) 2009-06-17 2022-03-17 3Shape A/S Focus scanning apparatus
US20210306617A1 (en) 2009-06-17 2021-09-30 3Shape A/S Focus scanning apparatus
US10595010B2 (en) 2009-06-17 2020-03-17 3Shape A/S Focus scanning apparatus
US20200169722A1 (en) 2009-06-17 2020-05-28 3Shape A/S Focus scanning apparatus
US20190124323A1 (en) 2009-06-17 2019-04-25 3Shape A/S Focus scanning apparatus
US11051002B2 (en) 2009-06-17 2021-06-29 3Shape A/S Focus scanning apparatus
US20210211638A1 (en) 2009-06-17 2021-07-08 3Shape A/S Focus scanning apparatus
US11076146B1 (en) 2009-06-17 2021-07-27 3Shape A/S Focus scanning apparatus
US9554692B2 (en) 2009-06-18 2017-01-31 EndoChoice Innovation Ctr. Ltd. Multi-camera endoscope
JP2009238245A (en) 2009-07-13 2009-10-15 Namco Bandai Games Inc Image generation system and information storage medium
WO2011011193A1 (en) 2009-07-21 2011-01-27 Dimensional Photonics International, Inc. Integrated display in a hand-held three-dimensional metrology system
US8547374B1 (en) 2009-07-24 2013-10-01 Lockheed Martin Corporation Detection and reconstruction of 3D objects with passive imaging sensors
US9084568B2 (en) 2009-08-05 2015-07-21 Telesystems Co., Ltd. Radiation imaging apparatus and imaging method using radiation
US9554857B2 (en) 2009-09-14 2017-01-31 Memorial Sloan-Kettering Cancer Center Apparatus, system and method for providing laser steering and focusing for incision, excision and ablation of tissue in minimally-invasive surgery
US8867820B2 (en) 2009-10-07 2014-10-21 Microsoft Corporation Systems and methods for removing a background of an image
WO2011047731A1 (en) 2009-10-22 2011-04-28 Tele Atlas B.V. Method for creating a mosaic image using masks
US20110125304A1 (en) 2009-11-24 2011-05-26 Sirona Dental Systems Gmbh Systems, methods, apparatuses, and computer-readable storage media for designing and manufacturing prosthetic dental items
EP2325771A2 (en) 2009-11-24 2011-05-25 Sirona Dental Systems GmbH Systems, methods, apparatuses, and computer-readable storage media for designing and manufacturing prosthetic dental items
US9208612B2 (en) 2010-02-12 2015-12-08 The University Of North Carolina At Chapel Hill Systems and methods that generate height map models for efficient three dimensional reconstruction from depth information
US20110200249A1 (en) 2010-02-17 2011-08-18 Harris Corporation Surface detection in images based on spatial data
WO2011120526A1 (en) 2010-03-30 2011-10-06 3Shape A/S Scanning of cavities with restricted accessibility
US9322646B2 (en) 2010-04-09 2016-04-26 The Trustees Of The Stevens Institute Of Technology Adaptive mechanism control and scanner positioning for improved three-dimensional laser scanning
US8260539B2 (en) 2010-05-12 2012-09-04 GM Global Technology Operations LLC Object and vehicle detection and tracking using 3-D laser rangefinder
US20110310449A1 (en) 2010-06-17 2011-12-22 Eun-Soo Kim Method for generating 3d video computer-generated hologram using look-up table and temporal redundancy and apparatus thereof
WO2012000511A1 (en) 2010-06-29 2012-01-05 3Shape A/S 2d image arrangement
US20130218530A1 (en) 2010-06-29 2013-08-22 3Shape A/S 2d image arrangement
WO2012007003A1 (en) 2010-07-12 2012-01-19 3Shape A/S 3D modeling of an object using textural features
US20130218531A1 (en) 2010-07-12 2013-08-22 3Shape A/S 3d modeling of an object using textural features
US8571281B2 (en) 2010-07-13 2013-10-29 Carestream Health, Inc. Dental shade mapping
US8208704B2 (en) 2010-07-13 2012-06-26 Carestream Health, Inc. Dental shade mapping
US9299192B2 (en) 2010-07-19 2016-03-29 Align Technology, Inc. Methods and systems for creating and interacting with three dimensional virtual models
US20130110469A1 (en) 2010-07-19 2013-05-02 Avi Kopelman Methods and systems for creating and interacting with three dimensional virtual models
US20120062557A1 (en) 2010-09-10 2012-03-15 Dimensional Photonics International, Inc. Systems and methods for processing and displaying intra-oral measurement data
US9566138B2 (en) * 2010-10-01 2017-02-14 Rune Fisker Modeling and manufacturing of dentures
US8526700B2 (en) 2010-10-06 2013-09-03 Robert E. Isaacs Imaging system and method for surgical and interventional medical procedures
US20120141949A1 (en) 2010-10-12 2012-06-07 Larry Bodony System and Apparatus for Haptically Enabled Three-Dimensional Scanning
US9185388B2 (en) 2010-11-03 2015-11-10 3Dmedia Corporation Methods, systems, and computer program products for creating three-dimensional video sequences
US9329675B2 (en) 2010-12-06 2016-05-03 3Shape A/S System with 3D user interface integration
USRE48221E1 (en) 2010-12-06 2020-09-22 3Shape A/S System with 3D user interface integration
WO2012076013A1 (en) 2010-12-06 2012-06-14 3Shape A/S System with 3d user interface integration
US20140022352A1 (en) 2010-12-21 2014-01-23 3Shape A/S Motion blur compensation
US20140022356A1 (en) 2010-12-21 2014-01-23 3Shape A/S Optical system in 3d focus scanner
WO2012083960A1 (en) 2010-12-22 2012-06-28 3Shape A/S System and method for scanning objects being modified
US20120179035A1 (en) 2011-01-07 2012-07-12 General Electric Company Medical device with motion sensing
EP2664272A1 (en) 2011-01-11 2013-11-20 Kabushiki Kaisya Advance Oral imaging and display system
US20120195471A1 (en) 2011-01-31 2012-08-02 Microsoft Corporation Moving Object Segmentation Using Depth Images
US9662188B2 (en) 2011-02-14 2017-05-30 Ivoclar Vivadent Ag Method for producing a dental restoration part and CAD/CAM device
WO2012115862A2 (en) 2011-02-22 2012-08-30 3M Innovative Properties Company Space carving in 3d data acquisition
US20130335417A1 (en) 2011-02-22 2013-12-19 3M Innovative Properties Company Space carving in 3d data acquisition
US20130158694A1 (en) 2011-03-18 2013-06-20 Natural Dental Implants Ag Integrated Support Device For Providing Temporary Primary Stability to Dental Implants and Prosthesis, and Related Methods
US8897526B2 (en) 2011-05-06 2014-11-25 Sirona Dental Systems Gmbh Method, system, and computer-readable medium for uncovering and planning an accurate dental preparation
US20140146142A1 (en) 2011-07-08 2014-05-29 François Duret Three-dimensional measuring device used in the dental field
US9629551B2 (en) 2011-07-15 2017-04-25 3Shape A/S Detection of a movable object when 3D scanning a rigid object
US10064553B2 (en) 2011-07-15 2018-09-04 3Shape A/S Detection of a movable object when 3D scanning a rigid object
WO2013010910A1 (en) 2011-07-15 2013-01-24 3Shape A/S Detection of a movable object when 3d scanning a rigid object
US20130034823A1 (en) 2011-08-02 2013-02-07 Rongguang Liang Adaptive illumination method and apparatus for dental shade matching
US9827076B2 (en) 2011-12-22 2017-11-28 3M Innovative Properties Company Method and system for making a dental restoration
EP2620733A1 (en) 2012-01-27 2013-07-31 Ivoclar Vivadent AG Dental device
WO2013122662A1 (en) 2012-02-13 2013-08-22 3M Innovative Properties Company Dental milling block containing individualized dental article and process of production
US20140377718A1 (en) 2012-02-13 2014-12-25 3M Innovative Properties Company Dental milling block containing individualized dental article and process of production
US20130244197A1 (en) 2012-03-16 2013-09-19 Soek Gam Tjioe Dental Shade Matching Device
US8848991B2 (en) 2012-03-16 2014-09-30 Soek Gam Tjioe Dental shade matching device
US8903746B2 (en) 2012-03-22 2014-12-02 Audrey Kudritskiy System and method for viewing, modifying, storing, and running artificial neural network components
US20130260340A1 (en) 2012-03-29 2013-10-03 3M Innovative Properties Company Powder for enhancing feature contrast for intraoral digital image scanning
US9456963B2 (en) 2012-12-18 2016-10-04 Dentca, Inc. Photo-curable resin compositions and method of using the same in three-dimensional printing for manufacturing artificial teeth and denture base
WO2014125037A1 (en) 2013-02-13 2014-08-21 3Shape A/S Focus scanning apparatus recording color
US20160022389A1 (en) 2013-02-13 2016-01-28 3Shape A/S Focus scanning apparatus recording color
US20140255878A1 (en) 2013-03-07 2014-09-11 A.Tron3D Gmbh Method for optical acquisition of the three-dimensional geometry of objects
US20160067018A1 (en) 2013-04-29 2016-03-10 3M Innovation Properties Company Method of capturing data from a patient's dentition and system for performing such method
EP2799032A1 (en) 2013-04-29 2014-11-05 3M Innovative Properties Company A method of capturing data from a patient's dentition and system for performing such method
US9844430B2 (en) * 2013-12-27 2017-12-19 James R. Glidewell Dental Ceramics, Inc. Apparatus and methods of making denture devices
US9707061B2 (en) * 2013-12-27 2017-07-18 James R. Glidewell Dental Ceramics, Inc. Apparatus and methods of making denture devices
US10111714B2 (en) 2014-01-27 2018-10-30 Align Technology, Inc. Adhesive objects for improving image registration of intraoral images
US10695151B2 (en) * 2014-02-07 2020-06-30 3Shape A/S Detecting tooth shade
US10010387B2 (en) 2014-02-07 2018-07-03 3Shape A/S Detecting tooth shade
US20180153664A1 (en) 2014-02-07 2018-06-07 3Shape A/S Detecting tooth shade
US9845745B2 (en) 2015-07-08 2017-12-19 Ford Global Technologies, Llc EVAP system with valve to improve canister purging
US10835361B2 (en) 2016-02-24 2020-11-17 3Shape A/S Detecting and monitoring development of a dental condition

Non-Patent Citations (226)

* Cited by examiner, † Cited by third party
Title
3Shape A/S Markman Heating Presentation, Case No. 1:18-886-LPS, Apr. 21, 2020, 104 pages.
3Shape A/S v. Align Technology, Inc., IPR2021-01383, Petition for Inter Partes Review, U.S. Pat. No. 10,728,519, Aug. 20, 2021, 112 pages.
Agrawal, Color and Shade Management in Esthetic Dentistry, Dec. 1, 2013, 120-127, vol. 3, Issue 3.
Ahn et al., "Development of Three-Dimensional Dental Scanning Apparatus Using Structured Illumination", Sensors, vol. 17, Issue 7, 1634 (IPR2018-00197, Ex. 2004) (IPR2018-00198, Ex. 2002), 2017, 9 pages.
Amended Complaint for Patent Infringement, 3SHAPE A/S v. Align Technology, Inc., Case No. 18-886-LPS, Aug. 30, 2019, 166 pages.
Answer, Affirmative Defenses, and Counterclaims of Align Technology, Inc, 3SHAPE A/S v. Align Technology, Inc.,C.A. No. 18-886-LPS, Oct. 21, 2019, 46 pages.
Atieh Mohammada. , "Accuracy Evlaution of Intral-Oral Optical Impressions: A Novel Approach", Thesis, University of North Carolina at Chapel Hill, 2016, 87 pages.
Baltzer et al.," The Determination of the Tooth Colors", Quintessenz Zahntechnik, 30(7), 2004, 726-740.
Bernardini et al., "High-Quality Texture Reconstruction from Multiple Scans", IEEE Transactions on Visualization and Computer Graphics, vol. 7, No. 4, Oct.-Dec. 2001, pp. 318-332.
Birnbaum et al., "Dental Impressions Using 3D Digital Scanners: Virtual Becomes Reality", Compendium of Continuing Education in Dentistry, vol. 29, No. 8, Oct. 2008, 18 pages.
Bornik et al., "A Hybrid User Interface for Manipulation of Volumetric Medical Data", 3D User Interfaces, 2006, pp. 29-36 (8 pages).
Borse et al., Tooth shade analysis and selection in prosthodontics: A systematic review and meta-analysis, Apr. 7, 2020, J Indian Prosthodont Soc, 20, 131-140.
Bowman et al., "3D User Interfaces: Theory and Practice", Addison-Wesley, 2004, pp. 96-101 (9 pages).
BOWMAN et at, "3D User Interfaces Theory and Practice § 4.1.1 "Input Device Characteristics" pp. 88-89; § 4.2.2 "2D Mice and Trackballs" pp. 91-92; § 4.8.2 "Input Device Taxonomies" pp. 128-132", Addison Wesley (IPR2018-00197, Ex. 2013), 2005, 20 pages.
Broadbent BH. , "A New X-Ray Technique and Its Application to Orthodontia", The Angie Orthodontist, vol. 1, No. 2, Apr. 1931, pp. 45-66.
Broadhurst et al., "A Probabilistic Framework for Space Carving", Proceedings Eighth IEEE International Conference on Computer Vision, vol. 1, 2001, 6 pages.
Callier et al., "Reconstructing Textured Meshes From Multiple Range+rgb Maps", 7th International Fall Workshop on Vision, Modeling, and Visualization, Nov. 2002,, 8 pages.
Chen et al., "A Volumetric Stereoetching Method: Application to Image-Based Modeling", IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No. PR00149). vol. 1, 1999, 6 pages.
Cherry RichardJ. , "New Techniques of Optical Microscopy and Microspectroscopy", The Macmillan Press Ltd., 1991, 3 pages.
Chu, Dental Color Matching Instrument and Systems. Review of Clinical and Research Aspects, Journal of Dentistry, 2000, e2-e16, vol. 38, Supplement 2.
Chua et al., "SonoDEX: 3D Space Management and Visualization of Ultrasound Data", International Congress Series, vol. 1281: (I PR2018-00197, Ex. 2006), May 2005, pp. 143-148.
Complaint, 3Shape A/S v. Carestream Dental, LLC, Civil Action No. 6:21-cv-1110, WDTX, Oct. 26, 2021, 59 pages.
Complaint, SShape A/S v. Carestream Dental LLC, Civil Action No. 6:21-cv-1110, 59 pages.
Corcodel et al., "Metameric effect between natural teeth and the shade tabs of a shade guide", May 11, 2010, European Journal of Oral Sciences, pp. 311-316.
Curriculum Vitae of Dr. Chandrajit L. Bajaj, (IPR2018-00197, Ex. 1004) (IPR2018-00198, Ex. 1004), 49 pages.
Curriculum Vitae of Dr. Ioannis A. Kakadiaris (56 pages).
Curriculum Vitae of Dr. James L. Mullins (13 pages).
Curriculum Vitae of Ravin Balakrishanan Ph.D, (IPR2018-00197, EX.2012), 30 pages.
Darrell et al., "Pyramid Based Depth from Focus", Proceedings CVPR '88: The Computer Society Conference on Computer Vision and Pattern Recognition, 1988; pp. 504-509.
Decision Denying Petitioner's Request for Rehearing in IPR20198-00198, U.S. Pat. No. 9,329,675, Dec. 4, 2018, 8 pages.
Declaration of Dr. Chandrajit Bajaj ("Bajaj Decl.") in support of Petition for Inter Partes Review, U.S. Patent No. RE48,221, 377 paages.
Declaration of Dr. Chandrajit L. Bajaj, (IPR2018-00197, Ex. 1003), Jul. 25, 2018, 142 pages.
Declaration of Dr. James L. Mullins (94 pages).
Declaration of Ioannis A. Kakadiaris (104 pages).
Declaration of Ravin Balakrishanan, (IPR2018-00197, EX.2011), 55 pages.
Defendant Align Technology, Inc.'s Initial Invalidity Contentions, 3Shape A/S v. Align Technology, Inc., C.A. No. 1:18-cv-00886-LPS, Nov. 21, 2017, 393 pages.
Defendant Align Technology, Inc.'s Stipulation of Invalidity Contentions, 3Shape A/S v. Align Technoiogv. Inc., C.A. No. 18-886-LPS, Exhibit 1053, Nov. 13, 2020, 3 pages.
Defendants' First Supplemental Invalidity Contentions (NDGA—Civil Action No. 1:22-cv-01829-WMR), dated Nov. 7, 2022 (101 pages).
Defendant's Identification of Invalidity References, 3SHAPE A/S et al. v. Align Technology, Inc., C. A. No. 1:18-cv-00886-LPS, 72 pages.
Defendants Initial Invalidity Contentions (NDGA—Civil Action No. 1:22-cv-01829-WMR), dated Sep. 12, 2022 (92 pages).
Deposition Transcript of Chandrajit Bajaj, Ph.D with Errata Sheet, (IPR2018-00197 Ex-2008), Jul. 25, 2018, 142 pages.
Deposition Transcript of Dr. Ravin Balakrishnan, Nov. 5, 2018, 101 pages.
DK priority document for Patent Application No. PA 2014 70066 (29 pages).
Douglas et al., "Intraoral determination of the tolerance of dentists for perceptibility and acceptability of shade mismatch", Apr. 1, 2007, Journal of Prosthetic Dentistry, 9(4), 200-208.
Eisert et al., "Automatic Reconstruction of Stationary 3-D Objects from Multiple Uncalibrated Camera Views", IEEE Transactions on Circuits and Systems for Video Technology, vol. 10, No. 2, Mar. 2000, pp. 261-277.
Eisert et al., "Multi-Hypothesis, Volumetric Reconstruction of 3-D Objects From Multiple Calibrated Camera Views", CASSP'99, Phoenix, USA, Mar. 1999, pp. 3509-3512.
Eisert Peter, "Reconstruction Volumetric 3D Models", 3D Video Communication: Algorithms, Concepts and Real-Time Systems in Human Centred Communication, 2001, pp. 1-20.
Elgammal Ahmed, "CS 534: Computer Vision Texture", 2003, 22 pages.
EP 2442720,"Interlocutory Decision in Opposition Proceedings", Jan. 15, 2019, 15 pages.
EP 2442720,"Notice of Opposition dated", Aug. 24, 2016, 5 pages.
EP 2442720,"Notice of Opposition", dated Aug. 24, 2016, 6 pages.
EP 2442720,"Notice of Opposition", dated May 22, 2017, 41 pages.
EP 2442720,"Notice of Opposition", dated May 24, 2017, 23 pages.
EPO Prosecution History dated Jun. 19, 2013, issued in the European Patent Application No. 11847582.1, 180 pages.
Exhibit 6—Latest Optical 3D Measuremental, Nov. 20, 2006, pp. 1-16.
File History (IPR2018-00197, Ex. 1002) (IPR2018-00198, Ex. 1002), U.S. Pat. No. 9,329,675, 625 pages.
File History, U.S. Patent No. RE48,221, 1004 pages.
Final Written Decision—Termination Decision Document from IPR2018-00197, U.S. Patent No. 9,329,675 B2, May 29, 2019, 64 pages.
First Office Action dated Apr. 3, 2015, issued in the corresponding Chinese Patent Application No. 201180066956.6, 13 pages.
First Office Action dated Dec. 2, 2016, issued in the corresponding Chinese Patent Application No. 201510098304.0, 15 pages including 8 pages of English Translation.
First Office Action dated Feb. 20, 2014, issued in the corresponding Chinese Patent Application No. CN201080027248.7, 22 pages including 13 pages of English Translation.
Fisher et al., "Dictionary of Computer Vision & Image Processing", Wiley, Second Edition, 2014, 386 pages.
Fisker et al., "Focus Scanning Apparatus", U.S. Appl. No. 61/187,744, filed Jun. 17, 2009, 90 pages.
Fisker et al., "Focus Scanning Apparatus", U.S. Appl. No. 61/231,118, filed Aug. 4, 2009, 127 pages.
Foley et al., "Introduction to Computer Graphics", Addison-Wesley, Chapter 2.2: Basic Interaction Handling, "Chapter 6: Viewing in 3D," and Chapter 8: input Devices, Interaction Techniques, and Interaction Tasks, 1994, 66 pages.
Forme ChristopherJ. , "3-D Scene Reconstruction From Multiple Photometric Images", PhD Thesis in Electrical and Computer Engineering at the University of Canterbury, Christchurch, NoW Zealand. Apr. 30, 2007, 179 pages.
Fraser et al., "Zoom-Dependent Camera Calibration in Digital Close-Range Photogrammetry", Photogrammetric Engineering & Remote Sensing, vol. 72, No. 9, Exhibit 1064, Sep. 2006, pp. 1017-1026.
Gao et al., "3D Shape Reconstruction of Teeth by Shadow Speckle Correlation Method", Optics and Lasers in Engineering, vol. 44, 2006, pp. 455-465.
Gehrung et al., "An Approach to Extract Moving Objects From MLS Data Using a Volumetric Background Representation", ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, vol. IV-1/W1 Jun. 2017, pp. 107-114.
Giammanco et al., "Using 3D Laser Scanning Technology to Create Digital Models of Hailstones", American Meteorological Society, Jul. 2017, pp. 1341-1347 (8 pages).
Gmitro et al., "Confocal Microscopy through a Fiber-Optic Imaging Bundle", Optics Letters, vol. 18, No. 8, Apr. 15, 1993, pp. 565-567 (4 pages).
Gonzalez et al., Digital Image Processing, Pearson Prentice Hall, 2008, Third Edition (977 pages).
Graetzel et al., "A Non-Contact Mouse for Surgeon-Computer Interaction", Technology and Health Care, vol. 12, No. 3, 2004, pp. 245-257.
Grant et al., "Glossary of Digital Dental Terms: American College of Prosthodontists", Journal of Prosthodontics, vol. 25, Suppl. 2, Oct. 2016, pp. S2-S9.
Guan et al., "Multi-view Occlusion Reasoning for Probabilistic Silhouette-Based Dynamic Scene Reconstruction". International Journal of Computer Vision, vol. 90, 2010, pp. 283-303.
Guehring Jens, "Dense 3D Surface Acquisition by Structured Light using off-the-Shelf Components", Proceedings SPIE 4309, Videometrics and Optical Methods for 3D Shape Measurement, Dec. 22, 2000, pp. 220-231 (13 pages).
Hajeer et al., "Current Products and Practices Applications of 3D Imagging in Orthodontics: Part II", Journal of Orthodontics, vol. 31, 2004, pp. 154-162.
Hale et al., "Measuring Free-Living Physical Activity in Adults with and Without Neurologic Dysfunction with a Triaxial Accelerometer", Archives of Physical Medicine and Rehabilitation, vol. 89, No. 9, Sep. 2008, pp. 1765-1771.
Havemann et al., "Seven Research Challenges of Generalized 3D Documents", IEEE Computer Graphics and Applications, vol. 27, No. 3, May-Jun. 2007, pp. 70-76.
Hearn et al., "Computer Graphics", 2d. Ed., Prentice Hall, Chapter 2: Overview of Graphics Systems. "Chapter 8: Graphical User Interfaces and Interactive Input Methods," and "Chapter 9: Three-Dimensional Concepts", 1994, 83 pages.
Horn et al., "Calculating the Reflectance Map", Applied Optics, vol. 18, No. 11, Jun. 1, 1979, pp. 1770-1779 (11 pages).
IEEE Xplore Search Results (4 pages), accessed Mar. 2, 2018; this document was made of record by the Examiner on Mar. 13, 2018, in the parent application (U.S. Appl. No. 15/117,078).
Institution Decision entered in IPR20198-00197, U.S. Pat. No. 9,329,675, May 30, 2018, 32 pages.
International Search Report (PCT/ISA/210) dated May 11, 2015, by the European Patent Office as the International Searching Authority for International Application No. PCT/EP2015/052537.
Introducing Wii MotionPlus, Nintendo's upcoming accessory for the revolutionary Wii Remote, Nintendo,, The Wayback Machine, Jul. 14, 2008, 2 pages.
Ireland et al., "3D Surface Imaging in Dentistry—What we are Looking at", British Dental Journal, vol. 205, No. 7, Oct. 11, 2008, pp. 387-392.
Ishikawa-Nagai et al., "Reproducibility of Tooth Color Gradation Using A Computer Color-Matching Technique Applied to Ceramic Restorations", J. Prosthetic Dentistry (Feb. 2005), pp. 129-137.
Jahne et al., "Handbook of Computer Vision and Applications", Sensors and Imaging, vol. 1, Academic Press, 1999, 657 pages.
Jahne, et al., "Handbook of Computer Vision and Applications", System and Applications, vol. 3, Academic press, 1999, 955 pages.
Jerald Jason, "The VR Book: Human-Centered Design for Virtual Reality", (IPR2018-00197 Ex-2014), 2016, 4 pages.
Jethwa Manish: "Efficient Volumetric Reconstruction from Multiple Calibrated Cameras", PhD Thesis in Electrical Engineering and Computer Science at MIT, Sep. 2004, 143 pages.
Johnstone et al., "Cameras Give Semiconductor Industry a Boost!", Nov. 7, 1985, 1 page.
JP 2014-23465 "Notification of Third Party Observations", Oct. 27, 2015, 31 pages.
JP 2014-234653,"Information Statement", English Translation, Jul. 28, 2016, 25 pages.
JP 2014-234653,"Notification of Information Statement", English Translation, Aug. 2, 2016, 2 pages.
Kang, "Three-Dimensional Lookup Table with Interpolation", Computational Color Technology, Chapter 9, 2006, pp. 151-159.
Karatas et al., "Three-Dimensional Imaging Techniques: A Literature Review", European Journal of Dentistry, vol. 8, Issue 1, Jan.-Mar. 2014, pp. 132-140.
Kaufmann Hannes, "Applications of Mixed Reality", Thesis, Vienna University of Technology, May 27, 2009, 95 pages.
Keating MichaelP. , "Geometric, Physical, and Visual Optics", 1988, 3 pages.
Lagouvardos et al., "Repeatability and Interdevice Reliability of Two Portable Color Selection Devices in Matching and Measuring Tooth Color", J. Prosthetic Dentistry (2007), 40-45.
Li et al., "Empty Space Skipping and Occlusion Clipping for Texture-based Volume Rendering", IEEE Visualization, 2003, 8 pages.
Litomisky et al., "Removing Moving Objects from Point Cloud Scenes", Advances in Depth Image Analysis and Applications, Jan. 2013, pp. 1-10.
Liu et al., "A Complete Statistical Inverse Ray Tracing Approach to Multi-View Stereo". Conferences, CVPR, 2011, pp. 913-920.
Logozzo et al., "Recent Advances in Dental Optics—Part I: 3D Intraoral Scanners for Restorative Denistry", Optics and Lasers in Engineering, vol. 54, Mar. 2014, pp. 203-221 (1-19).
Lovi Davidi. , "Incremental Free-Space Carving for Real-Time 3D Reconstruction", Master of Science Thesis in Computer Science at University of Alberta, 2011, 74 pages.
Mackinlay et al., "A Semantic Analysis of the Design Space of Input Devices", Human Computer Interaction, vol. 5, 1990, pp. 145-190.
Memorandum Order, 3Shape A/S v. Align Technology, Inc., C.A. No. 18-886-LPS, Exhibit 2022, Dec. 28, 2020, 3 pages.
Montes et al., "An Overview of BRDF Models", University of Granada, 2012, pp. 1-26.
Moran et al., "A Comparison of the Imaging Performance of High Resolution Ultrasound Scanners for Preclinical Imaging", Ultrasound in Medicine & Biology, vol. 37, No. 3, Mar. 2011, pp. 493-501.
Myers Brada. , "Graphical User Interface Programming", CRC Handbook of Computer Science and Engineering, 2d. Ed., Allen B. Tucker, Jan. 27, 2003, 30 pages.
Nasiri Steven, "A Critical Review of MEMS Gyroscopes Technology and Commercialization Status", InvenSense, 2005, 8 pages.
Netravali et al., "Digital Pictures: Representation, Compression, and Standards (Applications of Communications Theory)", 1988, 701 pages.
Nitschke et al., "Real-Time Space Carving Using Graphics Hardware", IEICE Transactions on Information and Systems, Aug. 2007, pp. 1175-1134 (11 pages).
Noguchi et al., "Microscopic Shape from Focus Using a Projected Illumination Pattern", Mathematical and Computer Modelling, vol. 24, No. 5/6, Sep. 1996, pp. 31-48.
Ogami Moria, "Exhibit 2-3D Imagery Handbook", First Edition, Feb. 20, 2006, pp. 1-4.
Öjelund et al., "Inter Partes Review Certificate", U.S. Pat. No. 9,329,675 K1, 2 pages.
Optical System for a Confocal Microscope, Certified English Translation of Application No. 01580/08. Oct. 6, 2008, 40 pages.
Order, Lipocine Inc. v. Clarus Therapeutics, Inc., C.A. No. 19-622 (WCB), Exhibit 1052, Nov. 12, 2020, 2 pages.
Paris et al., "A Surface Reconstruction Method using Global Graph Cut Optimization", International Journal of Computer Vision, vol. 66, No. 2, 2010, pp. 141-161.
Patent Owner's Preliminary Response to the Petition for Inter Partes Review in IPR2018-00198, U.S. Pat. No. 9,329,675, dated Mar. 3, 2018, 66 pages.
Patent Owner's Preliminary Response to the Petition for Inter Parties Review in IPR2018-00197, U.S. Pat. No. 9,329,675, dated Mar. 3, 2018, 66 pages.
Patent Owner's Response to the Petition for Inter Partes Review in IPR2018-00197, U.S. Pat. No. 9,329,675, Aug. 20, 2018, 57 pages.
Patent Owner's Submission of Demonstratives for Oral Argument in IPR2018-00197, U.S. Pat. No. 9,329,675, Jan. 31, 2019,, 42 pages.
PCT/DK2010/050148,"International Preliminary Report on Patentability Received", Jan. 5. 2012, 10 pages.
PCT/DK2010/050148,"International Search Report Received", dated Oct. 6, 2010, 5 pages.
PCT/DK2010/050148,"Written Opinion Received", Oct. 6, 2010, 8 pages.
PCT/DK2011/050461,"International Search Report PCT/ISA/210)", dated Feb. 22, 2012, 6 pages.
Petition for Inter Partes Review of U.S. Pat. No. 10,349,042, Case No. IPR2020-01089, Align Technology, Inc., Petitioner, v. 3Shape A/S, Patent Owner (76 pages).
Petition for Inter Partes Review of U.S. Pat. No. 10,695,151, dated Dec. 21, 2022 (87 pages).
Petitioner Align Technology, Inc.'s Demonstratives in IPR2018-00197, U.S. Pat. No. 9,329,675, Jan. 31, 2019., 30 pages.
Petitioner Align Technology, Inc.'s Reply to Patent Owner Response in IPR2018-00197, U.S. Pat. No. 9,329,675, Nov. 14, 2018,, 35 pages.
Petitioner's Align Technology, Inc's Request for Rehearing in IPR20198-00198, U.S. Pat. No. 9,329,675, Jun. 29, 2018, 14 pages.
Plaintiff and Counterclaim Defendant Align Technology, Inc.'s Stipulation Regarding IPR2022-00144 and IPR2022-00145,, Case No. 6:20-cv-00979 (W.D. Tex.), Dec. 16, 2021, 4 pages.
Pollard et al., "Change Detection in a 3-D World", IEEE Conference on Computer Vision and Pattern Recognition, Jun. 1, 2007, 6 pages.
Pulli et al., "Surface Reconstruction and Display from Range and Color Data", Graphical Models, vol. 62, Issue 3, 2000, pp. 165-201.
Pulli et al., "View-based Rendering: Visualizing Real Objects From Scanned Range and Color Data", Rendering Techniques'97, Springer, Vienna, 1997, 12 pages.
Record of Oral Hearing in IPR2018-00197, U.S. Pat. No, 9,329,675, Feb. 4, 2019, 67 pages.
Remondino et al., "Image-Based 3D Modelling: A Review", The Photogrammetric Record, vol. 21, No. 115, Sep. 2006, pp. 269-291.
Report and Recommendation, 3SHAPE A/S v. Align Technology, Inc., Case No. 18-886-LPS, May 6, 2020, 24 pages.
Richert et al., Intraoral Scanner Technologies: A Review to Make a Successful Impression, Journal of Healthcare Engineering, 2017, 9 pages.
Sato Yoichi, "Object Shape and Reflectance Modeling from Color Image Sequence", The Robotics Institute: Carnegie Mlellon University, Jan. 1997, 158 pages.
Savarese et al., "3D Reconstruction by Shadow Carving: Theory and Practical Evaluation", International Journal Computer Vision, vol. 71. No. 3, Mar. 2007, pp. 1-413.
Schendel et al., "3D Orthognathic Surgery Simulation Using Image Fusion", Seminars in Orthodontics, vol. 15, No. 1, Mar. 2009, pp. 48-56 (11 pages).
Schropp, Shade Matching Assisted by Digital Photography and Computer Software, Jan. 1, 2009, Journal of Prosthodontics, 235-241.
Second Office Action dated Nov. 18, 2015, issued in the corresponding Chinese Patent Application No. 201180066956.6, 27 pages including 16 pages of English Translation.
Sinescu et a., "Laser Beam Used in Dental Scanning for CAD/CAM Technologies", TMJ, vol. 57, No. 2-3, 2007, pp. 187-191 (6 pages).
Slabaugh et al., "Methods for Volumetric Reconstruction of Visual Scenes", International Journal of Computer Vision, vol. 57, 2004, pp. 179-199 (21 pages).
Slabaugh GregoryG. , "Novel Volumetric Scene Reconstruction Methods for New View Synthesis", PhD Thesis in Electrical and Computer Engineering at Georgia Institute of Technology, Nov. 2002, 209 pages.
Smith StevenW. , "Digital Signal Processing : A Practical Guide for Engineers and Scientists", 1998, 5 pages.
Smith WarrenJ. , "Modern Optical Engineering: The Design of Optical Systems", Third Edition, Exhibit 1065, 2000, 105 pages.
Spencer et al., "General Ray-Tracing Procedure", Journal of the Optical Society of America, vol. 52, No. 6, Jun. 1962, pp. 672-678.
Steele et al., "Bodies in Motion: Monitoring Daily Activity and Exercise with Motion Sensors in People with Chronic Pulmonary Disease", Journal of Rehabilitation Research & Development, vol. 40, No. 5, Suppl. 2, Oct. 2003, pp. 45-58.
Steinbach et al., "3-D Object Reconstruction Using Spatially Extended Voxels and Multi-Hypothesis Voxel Coloring", Proceedings 15th International Conference on Pattern Recognition, ICPR, vol. 1, 2000, 4 pages.
Tam, Dental Shade Matching Using a Digital Camera, Dec. 1, 2012, Journal of Dentistry, e3-e10, vol. 40, Supplement 2.
Tang et al., "Automatic Reconstruction of as-built Building Information Models from Laser-Scanned Point Clouds: A Review of Related Techniques", Automation in Construction, vol. 19, No. 7, Nov. 1, 2010, pp. 829-843.
Taxonomies of Input in Developing a Taxonomy of Input, (IPR2018-00197, Ex. 2010) Available at https:/iwww.billbuxton.com/input04. Taxonomies.pdf., Jan. 4, 2009, 16 pages.
Tiziani et al., "Theoretical Analysis of Confocal Microscopy with Microlenses". Applied Optics vol. 35, Issue 1, Jan. 1, 1996, pp. 120-125 (7 pages).
Toriwaki, et al., Fundamentals of Three-Dimensional Digital Image Processing, Springer, 2009, 278 pages.
Transcript of Alexander V. Sergienko, Ph.D., Align Technology. Inc. v. SShape A/S et al. Exhibit 1056, Jul. 16, 2021, 212 pages.
Transcript of Video Claim Construction Hearing, 3Shape A/S v. Align Technology, Inc., Case No. 18-886-LPS, Apr. 21, 2020, 137 pages.
Tsukizawa et al., "3D Digitization of a Hand-held Object with a Wearable Vision Sensor", International Workshop Computer Vision in Human-Computer Interaction, CVHCI 2004: Computer Vision in Human-Computer Interaction, 2004, pp. 129-141 (11 pages).
Tung et al., "The Repeatability of an Intraoral Dental Colorimeter", Dec. 1, 2002, J. Prosthetic Dentistry, 585-590.
Turner Daniel, "Hack: The Nintendo Wii", MIT Technology Review, Jul. 1, 2007, 3 pages.
U.S. Appl. No. 10/744,869, (IPR2018-00197, Ex. 2005), 69 pages.
U.S. Office Action for U.S. Appl. No. 15/888,764 (8 pages).
U.S. Pat. No. 10,349,042 "Declaration of Lambertus Hesselink, Ph.D.", Align Technology, Inc. v. 3Shape A/S, Case No. IPR2020-01087, Exhibit 1002, 193 pages.
U.S. Pat. No. 10,349,042 B1,"Final Written Decision", Align Technology, Inc. v. 3Shape A/S, IPR2020-01087, Jan. 19, 2022, 59 pages.
U.S. Pat. No. 10,349,042,"Declaration of Alexander Sergienko, Ph.D.". Align Technology, Inc. v. 3Shape A/S. Case No. IPR2020-01087, Exhibit 2019, 92 pages.
U.S. Pat. No. 10,349,042,"Declaration of Sylvia D. Hall-Ellis, Ph.D.", Align Technology, Inc. v. 3Shape A/S, Case No. IPR2020-01087, Exhibit 2029, 216 pages.
U.S. Pat. No. 10,349,042,"Deposition Transcript of Dr. Lambertus Hesselink", Align Technology, Inc. v. 3Shape A/S, IPR2020-01087, Exhibit 2023, Apr. 2, 2021, 82 pages.
U.S. Pat. No. 10,349,042,"Deposition Transcript of Dr. Lambertus Hesselink", Align Technology, Inc. v. 3Shape A/S, IPR2020-01087. Exhibit 2030, Sep. 10, 2021, 30 pages.
U.S. Pat. No. 10,349,042,"Patent Owner's Preliminary Response to the Petition for Inter Partes Review", Align Technology, Inc. v. 3Shape A/S, Case No. IPR2020-01088, Exhibit 1062, 78 pages.
U.S. Pat. No. 10,349,042,"Petition for Inter Partes Review", Align Technology, Inc., Petitioner, v. 3SHAPE A/S, Patent Owner, Case No. IPR202M1088, 81 pages.
U.S. Pat. No. 10,349,042,"Petition for Inter Partes Review", Align Technology. Inc., Petitioner v. 3SHAPE A/S, Patent Owner, Case No. IPR2020-01087, 89 pages.
U.S. Pat. No. 10,349,042,"Petition for Inter Partes Review", Align Technology. Inc., Petitioner, v. 3Shape NS, Patent Owner, Case No. IPR202001089, 119 pages.
U.S. Pat. No. 10,349,042,"Reply Declaration of Lambertus Hesselink", Align Technology, Inc. v. 3Shape A/S, Case No. IPR2020-01087, Exhibit 1057, 38 pages.
U.S. Pat. No. 61/420,138,"Ojelund Provisional", Dec. 6, 2010, 45 pages.
U.S. Pat. No. 8,363,228 B2,"Petition for Inter Partes Review", 3Shape A/S et al. v. Align, Technology, Inc., Case No. IPR2019-00157, Exhibit 1063, 81 pages.
U.S. Pat. No. 8,363,228,"Petition for Inter Partes Review", 3SHAPE A/S et al. v. Align Technology, Inc., Case No. IPR2019-00154, Nov. 10, 2018, 100 pages.
U.S. Pat. No. 9,329,675,"Decision Denying Institution of Inter Partes Review", Align Technology, Inc., Petitioner v. 3Shape A/S Patent Owner, Case IPR2018-00198, May 30, 2018, 15 pages.
U.S. Pat. No. 9,329,675,"Decision Institution of Inter Partes Review", Align Technology, Inc., Petitioner v 3Shape A/S Patent Owner, Case IPR2018-00197, May 30, 2018, 32 pages.
U.S. Pat. No. 9,329,675,"Decision Institution of Inter Parties Review", Align Technology, Inc., Petitioner v, 3Shape A/S Patent Owner, Case IPR2018-00197, May 30, 2018, 32 pages.
U.S. Pat. No. 9,329,675,"Declaration of Dr Chandrajit L Bajaj, Phd. In Support of Inter Partes Review", Align Technology, Inc., Petitioner v. 3Shape A/S Patent Owner, Case IPR2018-00197, 127 pages.
U.S. Pat. No. 9,329,675,"Dectaration of Dr. Chandrajit L. Bajaj. Ph.D. in Support of Inter Partes Review", Align Technology, Inc. Petitioner v. 3SHAPE A/S Patent Owner, Case IPR2018-00198, Ex. 1003, 123 pages.
U.S. Pat. No. 9,329,675,"Petition for Inter Partes Review", Align Technology Inc., Petitioner v. 3Shape A/S Patent Owner, Case IPR2018-00197, Nov. 22, 2017, 67 pages.
U.S. Pat. No. 9,329,675,"Petition for Inter Partes Review", Align Technology, Inc., Petitioner v. 3Shape A/S Patent Owner, Case IPR2018-00198, Nov. 22, 2017, 78 pages.
U.S. Pat. No. 9,329,675,"PTAB Trial Certificate Inter Partes Review Certificate", IPR Trial No. IPR2018-00197, Oct. 25, 2019, 2 pages.
U.S. Pat. No. 9,962,244,"Corrected Petition for Post-Grant Review", Align Technology, Inc. Petitioner, 3SHAPE A/S Patent Owner, Case No. PGR2018-00103, Oct. 30, 2018, 119 pages.
U.S. Pat. No. 9,962,244,"Declaration of Dr. Chandrajit L. Bajaj, Ph.D. in Support of Inter Partes Review", Align Technology, Inc. Petitioner, 3SHAPE A/S Patent Owner, Case No. IPR2019-00118, Nov. 5. 2018, 316 pages.
U.S. Pat. No. 9,962,244,"Declaration of Dr. Chandrajit L. Bajaj, Ph.D. In Support of Post-Grant Review", Align Technology, Inc. Petitioner v. 3SHAPE A/S Patent Owner, Case No. PGR2018-00104, Oct. 26, 2018, 318 pages.
U.S. Pat. No. 9,962,244,"Declaration of Dr. Chandrajit L. Bajaj. Ph.D. in Support of Post-Grant Review", Align Technology, Inc. Petitioner v, 3SHAPE A/S Patent Owner, Case Nos. PGR2018-00103, Oct. 30, 2018, 318 pages.
U.S. Pat. No. 9,962,244,"Declaration of Dr. Chandrajit L. Bajal, Ph.D. in Support of Post-Grant Review", Align Technology, Inc. Petitioner v. 3SHAPE A/S Patent Owner, Case Nos. PGR2018-00103, Oct. 30, 2018, 1272 pages.
U.S. Pat. No. 9,962,244,"Patent Owner's Preliminary Response to the Petition for Inter Partes Review", Align Technology, Inc. Petitioner v. 3Shape A/S, Patent Owner, Case No. IPR2019-00117, Mar. 4, 2019, 63 pages.
U.S. Pat. No. 9,962,244,"Patent Owner's Preliminary Response to the Petition for Inter Partes Review", Align Technology, Inc. Petitioner v. 3Shape A/S, Patent Owner, Case No. IPR2019-00118, Mar. 4, 2019, 62 pages.
U.S. Pat. No. 9,962,244,"Patent Owner's Preliminary Response to the Petition for Post-Grant Review", Align Technology, Inc. Petitioner v. 3Shape A/S Patent Owner, Case No. PGR2018-00103, Feb. 19, 2019, 64 pages.
U.S. Pat. No. 9,962,244,"Patent Owner's Preliminary Response to the Petition for Post-Grant Review", Align Technology, Inc. Petitioner v. 3Shape A/S, Patent Owner, Case No. PGR2018-00104, Feb. 19, 2019, 64 pages.
U.S. Pat. No. 9,962,244,"Petition for Inter Partes Review", Align Technology, Inc. Petitioner v. 3Shape A/S, Patent Owner, Case No. IPR2019-00118, Nov. 5, 2018, 93 pages.
U.S. Pat. No. 9,962,244,"Petition for Inter Partes Review", Align Technology, Inc., Petitioner V., 3Shape A/S, Patent Owner-Case No. IPR2019-00117, Nov. 5, 2018, 98 pages.
U.S. Pat. No. 9,962,244,"Petition for Post-Grant Review", Align Technology, Inc. , 3SHAPE A/S Patent Owner, Petitioner, Case No. PGR2018-00103, Oct. 30, 2016, 119 pages.
U.S. Pat. No. 9,962,244,"Petition for Post-Grant Review", Align Technology, Inc. Petitioner v. 3Shape A/S, Case No. PGR2018-00104, Oct. 26, 2018, 107 pages.
U.S. Pat. No. 9,962,244,"Second Corrected Declaration of Dr. Chandrajit L. Bajai, Ph.D. in Support of Post-Grant Review", Align Technology, Inc. Petitioner v. 3SHAPE A/S Patent Owner, Case No. PGR2018-00103, Oct. 30, 2018, 318 pages.
U.S. Pat. No. 9,962,244,"Second Corrected Petition for Post-Grant Review", Align Technology, Inc. Petitioner v. 3Shape A/S Patent Owner, Case No. PGR2018-00103, Oct. 30, 2018, 119 pages.
U.S. Pat. No. 9.676,430,"Declaration of Alexander Sergienko, Ph.D.", 3Shape A/S et al. v. Align Technology, Inc., Case No. IPR2020-01622, Exhibit 1061, 79 pages.
U.S. Pat. No. 9.962,244,"Declaration of Dr. Chandrajit L Bajaj, Ph.D. in Support of Inter Partes Review", Align Technology, Inc. Petitioner v. 3SHAPE A/S Patent Owner, Case No, IPR2019-00117, Nov. 5, 2018, 316 pages.
U.S. RE48,221,"Petition (1 of 2) for Inter Partes Review", Align Technology, Inc., Petitioner, 3SHAPE A/S, Patent Owner, Case No, IPR2022-00144, 99 pages.
U.S. RE48,221,"Petition (2 of 2) for Inter Partes Review", Align Technology, Inc., Petitioner, 3SHAPE A/S, Patent Owner, Case No. IPR2022-00145, 97 pages.
Vadher, et al., Basics of Color in Dentistry: A Review, Sep. 1, 2014, IOSR Journal of Dental and Medical Sciences, 13(9), 78-85.
Vedula et al., "Shape and Motion Carving in 6D", Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2000, pp. 1-7.
VIVID 910: Non-Contact 3-D Digitizer, www.minolta.com, 3 pages.
Vogt et al., "An AR System With intuitive User Interface for Manipulation and Visualization of 3D Medical Data", Studies in Health Technology and Informatics, vol. 98, 2004., pp. 397-403.
Watanabe et al., "Telecentric Optics for Constant-Magnification Imaging", Department of Computer Science, Columbia University, Sep. 1995, 22 pages.
Welch et al., "High-Performance Wide-Area Optical Tracking the HiBall Tracking System", Presence: Teleoperators and Virtual Environments, vol. 10, No. 1, Feb. 2001, pp. 1-22.
Welch et al., "Motion Tracking: No Silver Bullet, but a Respectable Arsenal", IEEE Computer Graphics and Applications, vol. 22, No. 6, Dec. 10, 2002, pp. 24-38.
Westphal et al., "Correction of Geometric and Refractive Image Distortions in Optical Coherence Tomorgraphy Applying Fermat's Principle", Optics Express, vol. 10, No. 9, May 6, 2002, pp. 397-404.
Wilson et al., "Confocal Microscopy by Aperture Correlation", Optics Letters vol. 21, Issue 23, 1996, pp. 1879-1881 (4 pages).
Wilson et al., "Dynamic Lens Compensation for Active Color Imaging and Constant Magnification Focusing", The Robotics Institute, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, Exhibit 2027, Nov. 1991. 52 pages.
Wilson et al., "Real-Time Three-Dimensional Imaging of Macroscorpic Structures", Journal of Microscopy, vol. 191, No. 2, Aug. 1998, pp. 116-118.
Written Opinion (PCT/ISA/237) dated May 11, 2015, by the European Patent Office as the International Searching Authority for International Application No. PCT/EP2015/052537.
Xia et al., "Three-Dimensional Virtual-Reality Surgical Planning and Soft-Tissue Prediction for Orthognathic Surgery", IEEE Transactions on Information Technology in Biomedicine. vol. 5, No. 2, Jun. 2001, pp. 97-107.
Xiao et al., "Efficient Partial-Surface Registration for 3D Objects", Computer Vision and Image Understanding, vol. 98, No. 2, 2005, pp. 271-294.
Yamany et al., "Free-Form Surface Registration Using Surface Signatures", The Proceedings of the Seventh IEEE International Conference on Computer Vision, vol. 2, 1999, 7 pages.
Yang et al., "Dealing with Textureless Regions and Specular Highlights—A Progressive Space Carving Scheme Using a Novel Photo-Consistency Measure", Proceedings of the Ninth IEEE International Conference on Computer Vision (ICCV 2003) 2-Volume Set, 2003, 9 pages.
Yoshida et al., "Intraoral Ultrasonic Scanning as a Diagnostic Aid", Journal of Cranio-Maxillofacial Surgery, vol. 15, 1987, pp. 306-311.
Yoshizawa Toru, "Handbook of Optical Metrology Principles and Application", Second Edition, Feb. 25, 2009, 15 pages.
Yuan et al., "Inferring 3D Volumetric Shape of Both Moving Objects and Static Background Observed by a Moving Camera", IEEE Conference on Computer Vision and Pattern Recognition, 2007, 8 pages.
Zhang et al., "A 3-dimensional Vision System for Dental Applications", Proceedings of the 29th Annual International, Aug. 23-26, 2007, pp. 3369-3372.

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11986325B2 (en) * 2019-06-13 2024-05-21 Osstemimplant Co., Ltd. Treatment information display device and method for displaying treatment history on image of teeth in accumulated manner

Also Published As

Publication number Publication date
US20170165038A1 (en) 2017-06-15
US11707347B2 (en) 2023-07-25
US11723759B2 (en) 2023-08-15
US10010387B2 (en) 2018-07-03
WO2015118120A1 (en) 2015-08-13
US20180153664A1 (en) 2018-06-07
US20220265403A1 (en) 2022-08-25
US20200352688A1 (en) 2020-11-12
US10695151B2 (en) 2020-06-30
US20220265402A1 (en) 2022-08-25

Similar Documents

Publication Publication Date Title
US11723759B2 (en) Detecting tooth shade
US11903779B2 (en) Dental preparation guide
CN109069237B (en) Method and computer readable medium for detecting and monitoring the development of a dental condition
US8144954B2 (en) Lighting compensated dynamic texture mapping of 3-D models
KR20210138652A (en) System and method for generating digital three-dimensional dental models
JP2022525088A (en) Dental shade matching for multiple anatomical areas
CN113039587B (en) Hybrid method for acquiring 3D data using intraoral scanner
KR102688481B1 (en) Method and apparatus for designing margin line of abutment model
EP4194804A1 (en) Method and device for acquiring three-dimensional data, and computer-readable storage medium storing program for performing method

Legal Events

Date Code Title Description
AS Assignment

Owner name: 3SHAPE A/S, DENMARK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ESBECH, BO;FISKER, RUNE;HENRIKSEN, LARS;AND OTHERS;SIGNING DATES FROM 20170330 TO 20170331;REEL/FRAME:052881/0218

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: AWAITING TC RESP., ISSUE FEE NOT PAID

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STCF Information on status: patent grant

Free format text: PATENTED CASE

STCF Information on status: patent grant

Free format text: PATENTED CASE