Nothing Special   »   [go: up one dir, main page]

EP3017428B1 - Ultrasonic imaging apparatus and control method thereof - Google Patents

Ultrasonic imaging apparatus and control method thereof Download PDF

Info

Publication number
EP3017428B1
EP3017428B1 EP14819625.6A EP14819625A EP3017428B1 EP 3017428 B1 EP3017428 B1 EP 3017428B1 EP 14819625 A EP14819625 A EP 14819625A EP 3017428 B1 EP3017428 B1 EP 3017428B1
Authority
EP
European Patent Office
Prior art keywords
sample
texel
ultrasonic
imaging apparatus
depth
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
EP14819625.6A
Other languages
German (de)
French (fr)
Other versions
EP3017428A4 (en
EP3017428A1 (en
Inventor
Young Ihn Kho
Hee Sae Lee
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Publication of EP3017428A1 publication Critical patent/EP3017428A1/en
Publication of EP3017428A4 publication Critical patent/EP3017428A4/en
Application granted granted Critical
Publication of EP3017428B1 publication Critical patent/EP3017428B1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/10Constructive solid geometry [CSG] using solid primitives, e.g. cylinders, cubes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/08Volume rendering
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/44Constructional features of the ultrasonic, sonic or infrasonic diagnostic device
    • A61B8/4405Device being mounted on a trolley
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/46Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
    • A61B8/461Displaying means of special interest
    • A61B8/463Displaying means of special interest characterised by displaying multiple images or images and diagnostic data on one display
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/46Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
    • A61B8/461Displaying means of special interest
    • A61B8/466Displaying means of special interest adapted to display 3D data
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/48Diagnostic techniques
    • A61B8/483Diagnostic techniques involving the acquisition of a 3D volume of data
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/52Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/5207Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of raw data to produce diagnostic data, e.g. for generating an image
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/54Control of the diagnostic device
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S15/00Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
    • G01S15/88Sonar systems specially adapted for specific applications
    • G01S15/89Sonar systems specially adapted for specific applications for mapping or imaging
    • G01S15/8906Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
    • G01S15/8993Three dimensional imaging systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/04Texture mapping
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/20Finite element generation, e.g. wire-frame surface description, tesselation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/20Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2210/00Indexing scheme for image generation or computer graphics
    • G06T2210/41Medical
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2215/00Indexing scheme for image rendering
    • G06T2215/16Using real world measurements to influence rendering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2219/00Indexing scheme for manipulating 3D models or images for computer graphics
    • G06T2219/20Indexing scheme for editing of 3D models
    • G06T2219/2012Colour editing, changing, or manipulating; Use of colour codes

Definitions

  • Embodiments of the present invention relate to an ultrasonic imaging apparatus and a control method thereof.
  • An ultrasonic imaging apparatus emits ultrasonic signals towards a specific part of a subject (e.g., a human body) from the skin surface of the subject, and receives ultrasonic signals (that is, ultrasonic echo signals) reflected from the specific part of the subject so as to non-invasively acquire section images about soft tissue or images about blood vessels based on information about the ultrasonic echo signals.
  • a subject e.g., a human body
  • ultrasonic signals that is, ultrasonic echo signals
  • the ultrasonic imaging apparatus is a low-priced apparatus having a small size, can display images in real time, and have high safety since there is no risk for patients to be exposed to radiation such as X-rays.
  • the ultrasonic imaging apparatus is widely used to diagnose the heart, breasts, abdomen, urinary organs, uterus, etc.
  • an ultrasonic imaging apparatus includes a main body to accommodate main components of the ultrasonic imaging apparatus, an ultrasonic collector to transmit and receive ultrasonic waves, a control panel including various switches and keys through which commands for manipulating the ultrasonic imaging apparatus are input, and a display to display the results of ultrasonic diagnosis as images.
  • a method in which the ultrasonic imaging apparatus performs ultrasonography on a subject is as follows. First, an operator holds the ultrasonic collector with his/her one hand to move the ultrasonic collector while keeping the ultrasonic collector in contact with the skin surface of a subject, and manipulates the control panel with the other hand to perform ultrasonography. An ultrasonic image obtained by the ultrasonography is displayed in real time through the display so that the operator can diagnose the state of the subject.
  • ultrasonic images obtained by ultrasonography are monochrome images due to the physical properties of the ultrasonic imaging apparatus, there is limitation in representing the inner structure of a subject realistically.
  • a method of producing 3-dimensional (3D) color ultrasonic images based on arbitrarily selected colors has been developed.
  • the method still has a problem that reality of images is low since the images are not based on the real colors of a subject.
  • the method results in lower reality than when producing images of the heart, breasts, etc.
  • US 2013/150719 relates to an ultrasound imaging system and method for ultrasound imaging.
  • the method includes generating a volume-rendered image from three-dimensional ultrasound data.
  • the volume-rendered image is colorized with at least two colors according to a depth-dependent color scheme.
  • the method includes displaying the volume-rendered image.
  • the method includes generating a planar image from the three-dimensional ultrasound data, where the planar image is colorized according to the same depth-dependent color scheme.
  • the method includes displaying the planar image.
  • EP 2 441 389 relates to an ultrasonic diagnostic apparatus to generate a three-dimensional tomographic image including: an offset calculating section which increases or decreases the brightness value of each voxel according to the brightness value of each voxel of a three-dimensional tomographic image volume data.
  • the amount of increase or decrease in the brightness value of each voxel of the offset calculating section is adjustable through a control panel, and a tomographic image volume rendering section generates the three-dimensional tomographic image on the basis of a three-dimensional tomographic image volume data in which the brightness value is offset by the offset calculating section.
  • an ultrasonic imaging apparatus includes: an ultrasonic collector configured to collect ultrasonic waves from an object; a volume data generator configured to generate volume data based on the ultrasonic waves; and an image processor configured to perform volume rendering on the volume data with reference to a texture image, wherein a translucency property and multi-layer tissue of the object are reflected to each texel of the texture image.
  • an ultrasonic imaging apparatus includes: an ultrasonic collector configured to collect ultrasonic waves from an object; a volume data generator configured to generate volume data based on the ultrasonic waves; and an image processor configured to perform volume rendering on the volume data with reference to a texture image, wherein a translucency property of the object is reflected to each texel of the texture image.
  • an ultrasonic imaging apparatus includes: an ultrasonic collector configured to collect ultrasonic waves from an object; a volume data generator configured to generate volume data based on the ultrasonic waves; and an image processor configured to perform volume rendering on the volume data with reference to a texture image, wherein multi-layer tissue of the object is reflected to each texel of the texture image.
  • a control method of an ultrasonic imaging apparatus includes: collecting ultrasonic waves from an object; generating volume data based on the ultrasonic waves; and performing volume rendering on the volume data with reference to a texture image, wherein a translucency property and multi-layer tissue of the object are reflected to each texel of the texture image.
  • a control method of an ultrasonic imaging apparatus includes: collecting ultrasonic waves from an object; generating volume data based on the ultrasonic waves; and performing volume rendering on the volume data with reference to a texture image, wherein a translucency property of the object is reflected to each texel of the texture image.
  • a control method of an ultrasonic imaging apparatus includes: collecting ultrasonic waves from an object; generating volume data based on the ultrasonic waves; and performing volume rendering on the volume data with reference to a texture image, wherein multi-layer tissue of the object is reflected to each texel of the texture image.
  • volume data By rendering volume data after reflecting a translucency property and multi-layer tissue of a subject to the volume data, it is possible to control and produce a realistic image such that it looks close to the real structure of the subject.
  • the ultrasonic imaging apparatus may display the fetus's figure realistically, which leads to improvement of emotional stability and mental satisfaction of fragment woman and her family.
  • FIG. 1 is a perspective view of an ultrasonic imaging apparatus according to an embodiment.
  • the ultrasonic imaging apparatus may include a main body 100, an ultrasonic collector 110, an input unit 150, a main display unit 160, and a sub display unit 170.
  • one or more female connectors 145 may be provided in one side of the main body 100.
  • a male connector connected to a cable 130 may be physically coupled with one of the female connectors 144.
  • a plurality of castors for moving the ultrasonic imaging apparatus may be provided.
  • the plurality of castors may be used to fix the ultrasonic imaging apparatus at a specific location and to move it in a specific direction.
  • the ultrasonic collector 110 contacts the skin surface of an object, and may transmit or receive ultrasonic waves. More specifically, the ultrasonic collector 110 functions to irradiate transmission signals (that is, ultrasonic signals) received from the main body 100 to the inside of the object, receive ultrasonic echo signals reflected from a specific part of the object, and transfer the ultrasonic echo signals to the main body 100.
  • the ultrasonic collector 110 may be connected to one end of the cable 130, and the other end of the cable 130 may be connected to the male connector 140.
  • the male connector 140 connected to the other end of the cable 130 may be physically coupled with one of the female connectors 145 of the main body 145.
  • the input unit 150 is used to receive commands for instructing operations of the ultrasonic imaging apparatus.
  • a mode selection command such as an Amplitude mode (A-mode), a Brightness mode (B-mode), or a Motion mode (M-mode), or an ultrasonography start command may be input through the input unit 150.
  • a command input through the input unit may be transferred to the main body 100 through wired/wireless communication.
  • the input unit 150 may include at least one of a keyboard, a foot switch, and a foot pedal.
  • the keyboard may be hardwarily implemented, and disposed on the upper part of the main body 100.
  • the keyboard may include at least one(s) of a switch, keys, a joystick, and a trackball.
  • the keyboard may be softwarily implemented as a graphic user interface (GUI).
  • GUI graphic user interface
  • the keyboard may be displayed through the sub display unit 170 or the main display unit 160.
  • the foot switch or the foot pedal may be provided in the lower part of the main body 100, and an operator may control operations of the ultrasonic imaging apparatus using the foot pedal.
  • An ultrasonic collector holder 120 for accommodating the ultrasonic collector 110 may be provided to be adjacent to the input unit 150. There may be provided one or more ultrasonic collector holders. The operator may put the ultrasonic collector 110 into the ultrasonic collector holder 120 to safely keep the ultrasonic collector 110 when he/she does not use the ultrasonic imaging apparatus.
  • the sub display unit 170 may be installed in the main body 100.
  • FIG. 1 shows a case in which the sub display unit 170 is provided on the upper part of the input unit 150.
  • the sub display unit 170 may display applications related to operations of the ultrasonic imaging apparatus.
  • the sub display unit 170 may display menus or guidance for ultrasonography.
  • the sub display unit 170 may be implemented as a Cathode Ray Tube (CRT) or a Liquid Crystal Display (LCD).
  • the main display unit 160 may be also installed in the main body 100.
  • FIG. 1 shows a case in which the main display unit 160 is positioned above the sub display unit 170.
  • the main display unit 160 may display ultrasonic images acquired by ultrasonography.
  • the main display 160 may be, like the sub display 170, implemented as a CRT or a LCD.
  • the main display unit 160 is integrated with the main body 100, however, the main display unit 160 may be separated from the main body 100.
  • the ultrasonic imaging apparatus includes both the main display unit 160 and the sub display unit 170, however, the sub display unit 170 may be omitted. In this case, applications, menus, etc., which are displayed through the sub display unit 170, may be display through the main display unit 160.
  • FIG. 2 is a block diagram of a control configuration of an ultrasonic imaging apparatus according to an embodiment.
  • the ultrasonic collector 110 includes a plurality of ultrasonic transducers 114 to generate ultrasonic waves according to alternating current applied from a power source 112, to irradiate the ultrasonic waves to an object, to receive echo ultrasonic waves reflected from a target part of the object, and to convert the received echo ultrasonic waves into electrical signals.
  • the power source 112 may be an external power supply, or a power storage unit included in the ultrasonic imaging apparatus.
  • Each ultrasonic transducer 114 may be a magnetostrictive ultrasonic transducer using the magnetostrictive effect of a magnetic material, a piezoelectric ultrasonic transducer using the piezoelectric effect of a piezoelectric material, or a capacitive micromachined ultrasonic transducer (CMUT) that transmits and receives ultrasonic waves using vibration of several hundreds or thousands of micromachined thin films.
  • CMUT capacitive micromachined ultrasonic transducer
  • each ultrasonic transducer 114 When alternating current from the power source 112 is applied to each ultrasonic transducer 114, the piezoelectric vibrator or thin film of the ultrasonic transducer 114 vibrates, thus generating ultrasonic waves.
  • the generated ultrasonic waves are irradiated to an object, for example, to the inside of a human body.
  • the irradiated ultrasonic waves are reflected by one or more target areas that may be located at different depths from the surface of the object.
  • the ultrasonic transducers 114 receive ultrasonic echo signals reflected from the target areas, and convert the received ultrasonic echo signals into electrical signals to obtain a plurality of received signals.
  • the received signals are transferred to the main body 100 through a wired/wireless communication network. Since the ultrasonic collector 110 receives the ultrasonic echo signals through a plurality of channels, the plurality of received signals are also transferred to the main body 100 through a plurality of channels.
  • the main body 100 may include a beamformer 210, a volume data generator 220, and an image processor 230. In another embodiment, one or more components of the above-mentioned components may be omitted.
  • the beamformer 210 may perform beamforming based on the plurality of received signals. Beamforming is operation of focusing a plurality of received signals received through a plurality of channels to acquire a proper ultrasonic image about the inside of an object.
  • the beamformer 210 corrects time differences of a plurality of received signals, caused by distance differences between the respective transducers 114 and a target area of the object. Then, the beamformer 210 enhances a plurality of received signals corresponding to specific channels, or relatively attenuates a plurality of received signals corresponding to other channels, thereby focusing the plurality of received signals. At this time, the beamformer 210 may enhance or attenuate a specific received signal by selectively allocating predetermined weights to a plurality of received signals received through individual channels.
  • the beamformer 210 may focus a plurality of received signals collected by the ultrasonic collector 110, for each frame, in consideration of the locations and focusing points of the transducers 114 of the ultrasonic collector 110.
  • the beamforming that is performed by the beamformer 210 may be one of data-independent beamforming and adaptive beamforming.
  • the volume data generator 220 generates ultrasonic image data based on the signals focused by the beamformer 210. If the beamformer 210 focuses a plurality of received signals for each of a plurality of frames, the volume data generator 220 generates a plurality of pieces of ultrasonic image data respectively corresponding to the respective frames based on a plurality of focused signals.
  • the ultrasonic image data may be 3D ultrasonic image data for the object, that is, volume data for the object.
  • the image processor 230 may receive the volume data for the object from the volume data generator 230. Then, the image processor 230 performs volume rendering on the volume data, and transfers the results of the volume rendering to the main display unit 160 so that the main display unit 160 can output the results of the volume rendering as images.
  • the image processor 230 may volume-render the 3D volume data using one of volume rendering methods well-known in the art.
  • volume rendering may be classified into surface rendering and direct volume rendering.
  • the surface rendering is to extract surface information from volume data based on predetermined scalar values and amounts of spatial changes, to convert the surface information into a geometric factor, such as a polygon or a surface patch, and then to apply a conventional rendering technique to the geometric factor.
  • a geometric factor such as a polygon or a surface patch
  • Examples of the surface rendering are a marching cubes algorithm and a dividing cubes algorithm.
  • the direct volume rendering is to directly render volume data without converting volume data into a geometric factor.
  • the direct volume rendering is useful to represent a translucent structure since it can visualize the inside of an object as it is.
  • the direct volume rendering may be classified into an object-order method and an image-order method according to a way of approaching volume data.
  • the image-order method is to sequentially decide pixel values of an image.
  • An example of the image-order method is volume ray casting.
  • a virtual ray is irradiated from a specific viewpoint toward a predetermined pixel of a display screen, and then voxels through which the virtual ray has been transmitted from among voxels of volume data are detected. Then, brightness values of the detected voxels are accumulated to decide a brightness value of the corresponding pixel of the display screen.
  • an average value of the detected voxels may be decided as a brightness value of the corresponding pixel of the display screen.
  • a weighted average value of the detected voxels may be decided as a brightness value of the corresponding pixel of the display screen.
  • the object-order method is to directly project volume data on an image.
  • An example of the object-order method is splatting. According to splatting, by aligning the center of a reconfigured kernel with a point at which a voxel is projected on a plane to be observed, and then causing peripheral pixels to distribute values to the voxel, a brightness value of the corresponding pixel is decided.
  • a 3D ultrasonic image may be produced.
  • Each pixel of the 3D ultrasonic image may have a brightness value and a color value. The following description is given under an assumption that the volume ray casting of the above-mentioned volume rendering methods is used.
  • the storage unit 240 may store data or algorithms required for operations of the ultrasonic imaging apparatus 200.
  • the storage unit 240 may store an algorithm for rendering volume data, an algorithm for calculating the location of a sample with respect to a viewpoint, and an algorithm for calculating a depth from a surface to each sample.
  • the storage unit 240 may be Read Only Memory (ROM), Random Access Memory (RAM), Programmable Read Only Memory (PROM), Erasable Programmable Read Only Memory (EPROM), a non-volatile memory device such as flash memory, a volatile memory device such as Random Access Memory (RAM), or an optical disk.
  • ROM Read Only Memory
  • RAM Random Access Memory
  • PROM Programmable Read Only Memory
  • EPROM Erasable Programmable Read Only Memory
  • a non-volatile memory device such as flash memory
  • RAM Random Access Memory
  • RAM Random Access Memory
  • optical disk an optical disk.
  • the storage unit 280 is not limited to these, and may be any other storage device well-known in the art.
  • the main body 100 of the ultrasonic imaging apparatus includes the beamformer 210, the volume data generator 220, the image processor 230, and the storage unit 240 has been described above with reference to FIGS. 1 and 2 .
  • the main body 100 does not need to include all the above-mentioned components. In other words, some of the components may be included in another component, for example, in the ultrasonic collector 110.
  • the ultrasonic collector 110 may include the beamformer 210 to produce an ultrasonic image based on electrical signals, and to correct the produced ultrasonic image.
  • the above description relates to the appearance and control configuration of an ultrasonic imaging apparatus according to an embodiment.
  • FIG. 3 is a block diagram of a control configuration of the image processor 230 according to an embodiment.
  • the image processor 230 may include a texture image storage unit 232, a sampling unit 234, a calculator 236, and a data accumulating unit 238.
  • the texture image storage unit 232 may store texture images that are referred to upon volume rendering.
  • the texture image storage unit 232 may be implemented as the same type of device as the storage unit 240 since the texture image storage unit 232 also functions to store data.
  • the texture image storage unit 232 may store one or more texture images.
  • a texture image that is to be stored in the texture image storage unit 232 may reflect a translucency property or multi-layer tissue of an object.
  • a translucency property or multi-layer tissue of an object will be described.
  • a phenomenon in which light is transmitted through a certain object without being absorbed in or reflected from the object is called transmission.
  • a degree of light transmission greatly depends on the properties of an object through which the light is transmitted.
  • a transparent material such as glass transmits a major part of incident light although it reflects a small amount of the incident light and absorbs a small amount of the incident light.
  • a degree at which light is transmitted through a material is determined as transparency of the material, and a material through which light is 100% transmitted can be considered to have transparency. Meanwhile, a material through which no light is transmitted is considered as an opaque material. If a certain material transmits a small amount of light, and scatters light according to its surface structure and properties to lower transparency, the material is considered to have a translucency property.
  • An object such as the fetus having soft tissue can be considered to have a translucency property since a constant amount of light can be transmitted through the object. Accordingly, such an object is strongly dependent on an effect due to a translucency property since a part of light irradiated from a light source is transmitted deeply to tissue.
  • an effect due to light transmission through the object must be considered even when a light source is positioned behind the object, as well as when a light source is positioned in front of the object.
  • a translucency property can be represented through calculations under an assumption that a part of incident light is transmitted through the inside of an object while the remaining part of the incident light is reflected from the surface of the object.
  • FIG. 4 illustrates light paths when a light source is located to face a viewpoint with an object in between.
  • no path of light reflected from the object is shown since the reflected light has no influence in observing the object at the viewpoint.
  • a path of light passing a point df1 at which light from a light source enters the inside of the object, and a point do1 at which the light transmitted through the inside of the object is escaped from the object is described below.
  • the light is transmitted through the inside of the object by a distance of S1.
  • a distance by which the light is transmitted through the inside of the object depends on a path of the light.
  • a distance S1 by which light is transmitted through the lower part of the object is longer than a distance S2 by which light is transmitted through the upper part of the object (S1>S2).
  • a part of light incident to an object is absorbed in the object when the light is transmitted through the inside of the object, so long as the object is not a completely transparent material. Accordingly, as a distance by which light is transmitted through the inside of an object is longer, the more amount of light is absorbed in the object. Therefore, in the case of FIG. 4 , when the object is seen from the viewpoint, the point do1 is observed darker than the point do2.
  • one axis of a texture image that is to be stored in the texture image storage unit 232 may be configured to represent a depth from a surface of an object which a light source faces.
  • the translucency property of an object With the translucency property of an object, multi-layer tissue of the object must be considered.
  • Light transmitted through a medium is influenced by the kind of the medium. Specifically, since different kinds of mediums absorb different wavelengths of light, the color of an object shown from a viewpoint depends on a depth to which light has been transmitted through the object. Accordingly, when an object is shown from a viewpoint, the effect of light transmitted through different material layers of the object, that is, the colors of light transmitted through the different material layers of the object are mixed and visually shown.
  • an object is a human body
  • multi-layer tissue of the human body can be considered.
  • the skin of a human body is not a single material and consists of several different kinds of layers. If the skin of the human body is divided into a horny layer, an epidermis layer, a dermis layer, and a panniculus layer, the object may be represented in a different way from an object having a single layer since the respective layers have different properties.
  • the multi-layer tissue of the object is considered upon rendering in order to improve reality.
  • volume rendering since the inside of an object as well as the surface of the object must be rendered, a method of deciding color values in consideration of multi-layer tissue of an object can improve the reality of rendering.
  • a depth at which a sample is located from a surface with respect to a viewpoint may be calculated.
  • a color value in consideration of the depth of a sample upon rendering reality can be improved.
  • one axis of the texture image may be configured to represent a depth from a surface with respect to a viewpoint.
  • FIG. 5 illustrates an example of a texture image including information about a translucency property and multi-layer tissue of an object (e.g., the fetus).
  • a texture image which is shown as black and white in FIG. 5 , is assumed to be in color since the present disclosure is implemented using a color texture image.
  • the texture image storage unit 232 may store 2-dimensional (2D) texture images.
  • Each of the 2D texture images may be configured with a group of texels each having a brightness value or a color value.
  • the brightness or color values of the texels may be referred to in order to improve reality.
  • the texel is a minimum unit constructing text image data.
  • a texture image may be configured such that the horizontal axis represents the depths of samples from a surface with respect to a viewpoint, and the vertical axis represents the depths of the samples from a surface with respect to a light source.
  • the depth (the horizontal axis of a texture image) of a sample from a surface with respect to a viewpoint is used to determine what layer of the corresponding object the sample is located in, and accordingly, the horizontal axis of the texture image may become criteria to decide a color value of a texel.
  • the depth (the vertical axis of a texture image) of a sample from a surface with respect to a light source is used to determine a depth to which light has been transmitted through the corresponding object, and accordingly, the vertical axis of the texture image may become criteria to decide a brightness value of a texel.
  • the sampling unit 234 receives volume data of an object from the volume data generator 220 (see FIG. 2 ). The sampling unit 234 performs sampling for volume rendering on the received volume data.
  • axes of a display screen may be set to axes of an orthogonal coordinates system, and the locations of samples may be represented as vectors. Thereby, distances between samples and a surface may be obtained through a simple mathematical calculation.
  • FIG. 6 is a view for describing a process of rearranging the position of an object with respect to a viewpoint upon sampling, according to an embodiment.
  • FIG. 6 shows an example of an arrangement of volume data samples, having a cube shape in which each side has scalar values of N voxels.
  • a display screen is configured with MxN pixels, wherein M is the number of pixels arranged on the horizontal axis of the display screen, and N is the number of pixels arranged on the vertical axis of the display screen.
  • Each light beam is projected to a pixel.
  • the z-axis coordinate w of the vector U increases in proportion to a sampling distance and the number of times of sampling.
  • Setting values for sampling that is performed by the sampling unit 234 may be input by a user through the input unit 150. For example, a sampling distance, the number of times of sampling, and a point at which sampling terminates may be set to desired values by a user.
  • the calculator 236 receives the location of each sample, acquired by sampling, from the sampling unit 234.
  • the calculator 236 functions to calculate a distance of the sample to the surface, using the received location of the sample.
  • the calculator 236 may be hardwarily implemented to enable parallel processing using a plurality of pipelines.
  • the calculator 236 may perform a calculation for detecting a texel corresponding to each sample in the direction of the light beam passing through the display screen.
  • the horizontal axis of the texture image may be set to a depth from a surface with respect to a viewpoint
  • the vertical axis of the texture image may be set to a depth from a surface with respect to a light source.
  • FIG. 7 is a view for describing a method of calculating the location of a texel corresponding to each sample, according to an embodiment.
  • a depth X of the corresponding sample from a surface with respect to a viewpoint is calculated.
  • a depth Y of the corresponding sample from a surface with respect to a light source is calculated.
  • a texel corresponding to the sample may be detected from the texture image of FIG. 6 .
  • the calculator 236 may transfer the color value and brightness value of the detected texel to the data accumulating unit 238.
  • a method of calculating the depths X and Y will be described in detail with reference to FIGS. 8 to 11 .
  • FIGS. 8A , 8B, and 8C are views for describing a method of calculating a depth X from a surface to each sample with respect to a viewpoint, according to an embodiment.
  • FIG. 8A when a virtual light beam is projected to an object from a viewpoint, volume data is sampled in the direction of the light beam. In order to find a surface on the path of the light beam, sampling from the viewpoint continues to be performed, and if it is found a first sample which is not an empty space, the location of the found sample is estimated as the location of a surface. If an empty space is found during continuous sampling, the location of a sample which is not an empty space and is first found is again estimated as the location of a surface.
  • a gradient vector of the estimated surface is calculated.
  • the gradient vector of the surface is calculated using texel values of peripheral samples on x-, y-, and z-axis.
  • the following Equation (1) is used to calculate an x component of a gradient vector of a surface.
  • a vector represents the location of a sample
  • f represents a voxel value of the sample
  • h represents a unit distance between samples, wherein u, v, and w are x-, y-, and z-axis coordinates.
  • g x U ⁇ u v w f U ⁇ u + 1 , v , w ⁇ f U ⁇ u ⁇ 1 , v , w 2 h
  • a plane in which the normal vector is the gradient vector of the surface is obtained.
  • the plane is a tangential plane of the object at a point corresponding to the surface.
  • a distance from the plane to the sample of which a depth has to be calculated is calculated, and the distance becomes a depth X from the surface to the sample with respect to the viewpoint.
  • FIG. 9 is a flowchart of a method of calculating a depth X from an object's surface to each sample with respect to a viewpoint, according to an embodiment.
  • the depth X can be calculated by obtaining a tangential plane at a point corresponding to the surface, and calculating a distance from the tangential plane to the sample.
  • sampling is performed along a light beam from a viewpoint (300).
  • the number of times of sampling increases in proportion to a degree of reality of volume rendering. The reason is because as the number of times of sampling is more, the more accurate location of a surface can be found.
  • the number of times of sampling may be set to a value input by a user through the input unit 150 (see FIG. 2 ). Sampling may be performed simultaneously with or separately from the following process, and in the following description, it is assumed that sampling is performed simultaneously with the following process.
  • voxel values of samples are checked in order from the viewpoint (310). Then, it is determined whether each sample is an empty space based on the voxel values (320). Whether each sample is an empty space may be determined using an Opacity Transfer Function (OTF) that defines the relationship between density values and transparency among the voxel values. Since a method of determining an empty space using an OTF is well-known in the art, further descriptions will be omitted.
  • OTF Opacity Transfer Function
  • a sample which is not an empty space is checked (325). If a sample which is not an empty space is found, the location values of the sample are acquired (330). The sample whose location values have been acquired is estimated as a surface. The reason is because a sample which is not an empty space and is first found is a first sample that exists in the inside of the object in the direction of the light beam.
  • operation for calculating a depth X of the sample is performed.
  • the depth X of the sample is calculated using the location values of the sample.
  • operation of calculating a gradient vector at the location of the sample is performed (340).
  • the gradient vector is calculated using peripheral values on the x-, y-, and z-axis of the sample. Since a method of calculating the gradient vector has been described above with reference to FIG. 8B , further descriptions will be omitted.
  • a plane on which the location of the sample estimated as the surface exists and in which the normal vector is the gradient vector of the surface is obtained (350).
  • the plane is a tangential plane of the object at a point corresponding to the surface since the plane meets the object only at the point of the surface and is perpendicular to the object.
  • the depth X of the sample from the surface with respect to the viewpoint may be acquired (360).
  • the depth X may be used to detect a texel on the horizontal axis of the texture image of FIG. 6 .
  • FIGS. 10A, 10B, and 10C are views for describing a method of calculating a depth Y from a surface to each sample with respect to a light source, according to an embodiment.
  • a light beam advancing from a sample of which a depth has to be calculated toward a light source is generated. This is aimed at calculating a distance by which light irradiated from a light source is transmitted through a medium until the light arrives at the corresponding sample to reflect the translucency property of the corresponding object. Since light reflected from the surface of the object has no influence in deciding a brightness value of the sample, only light advancing straightly is considered.
  • volume data is sampled at regular time intervals along the generated light beam. At this time, sampling is performed along the traveling direction of the light beam, starting from the sample of which the depth has to be calculated, so that brightness values among voxel values of samples are checked. Since a brightness value becomes greater as light irradiated from a light source is transmitted through a medium to the deeper depth, the brightness values are checked to detect a sample on the surface of a medium. The brightness values acquired in order of sampling are compared to a predetermined threshold value.
  • the previous sample may be estimated as a surface. Then, a distance between the surface and the sample of which the depth has to be calculated is calculated, so that a depth Y from the surface to the sample can be calculated.
  • FIG. 11 is a flowchart of a method of calculating a depth Y from an object's surface to each sample with respect to a light source, according to an embodiment.
  • the depth Y can be calculated based on a fact that the strength of light varies depending on a distance by which the light is transmitted through a medium until the light arrives at a sample.
  • a light beam advancing from a sample of which a depth has to be calculated toward a light source is generated (400). This is aimed at setting a path of light arriving at the corresponding sample from an actual light source, and easily calculating a distance by which the light advancing along the path is transmitted through a medium.
  • sampling is performed at regular time intervals along the traveling direction of the generated light beam (410).
  • the number of times of sampling increases in proportion to a degree of reality of volume rendering. The reason is because as the number of times of sampling is more, the more accurate location of a surface can be found.
  • the number of times of sampling may be set to a value input by a user through the input unit 150 (see FIG. 2 ). Sampling may be performed simultaneously with or separately from the following process, and in the following description, it is assumed that sampling is performed simultaneously with the following process.
  • brightness values among voxel values of samples are acquired in order, starting from a sample located most distant from the light source (420), and the brightness values are compared to a predetermined threshold value (430). By comparing the brightness values to the predetermined threshold value, it can be determined whether a sample is included in the inside of the object.
  • the predetermined threshold value may be set to an appropriate value within an error range of a brightness value when light is not transmitted through any medium.
  • the brightness value of a sample is compared to the predetermined threshold value, if the brightness value of the sample is greater than the predetermined threshold value, the corresponding sample is determined as a sample located in the inside of the object, and then, the next sample is compared to the predetermined threshold value (435). Meanwhile, if the brightness value of the sample is smaller than the predetermined threshold value, the sample is determined as a sample located outside the object.
  • the previous sample is estimated as a surface of the object. Then, the location values of the sample are acquired (440). The reason is because if a sample located outside the object is found after successive samples located in the inside of the object are found, the previous sample of the sample located outside the object is a final sample located in the inside of the object in the traveling direction of the light beam.
  • the depth Y may be used to detect a texel on the vertical axis of the texture image of FIG. 6 .
  • the calculator 236 may detect a texel corresponding to the sample from a texture image stored in the texture image storage unit 232 based on the resultant depth value (x, y). Then, a brightness value or a color value is detected from the corresponding texel.
  • the data accumulating unit 238 receives the brightness value or the color value of the texel detected by the calculator 236.
  • the brightness values or color values of texels corresponding to samples acquired along the path of a light beam passing through a display pixel with respect to a viewpoint are all accumulated in the data accumulating unit 238.
  • the brightness or color of the display pixel through which the light beam passes is decided.
  • FIG. 12 is a flowchart of a rendering algorithm for a volume rendering method, according to an embodiment.
  • sampling is performed at regular time intervals in the direction of a light beam from a viewpoint (500). At this time, as sampling is performed at shorter time intervals, more samples are subject to a calculation to decide a brightness value and a color value of each display pixel. Accordingly, since more brightness values and color values of texels are acquired, a more amount of data is accumulated to obtain the results of rendering having high reality.
  • a sampling interval may be set to a value input by a user through the input unit 150.
  • voxel values of samples are checked in order, starting from a sample located closest to a viewpoint (510). Then, it is determined whether each sample is an empty space (520). Whether the sample is an empty space may be determined using the OTF which is well-known in the art. If it is determined that a sample is an empty space based on the voxel value of the sample, it is determined whether the next sample is an empty space (525). The reason is because a sample corresponding to an empty space has no information about an object.
  • the location values of the sample are acquired (530), and a calculation for detecting a texel corresponding to the sample from a texture image is performed. To do the calculation, a depth of the sample from a surface of the object with respect to a light source is calculated (540), and a depth of the sample from a surface of the object with respect to the viewpoint is calculated (550).
  • the depth of the sample from the surface of the object with respect to the light source is calculated by calculating a distance by which a light beam generated from the sample passes through a medium until the light beam arrives at the light source.
  • the depth of the sample from the surface of the object with respect to the viewpoint is calculated by estimating a sample which is not an empty space and is first found as a surface, and calculating a distance between a tangential plane of the object at a point corresponding to the surface and the sample of which the depth has to be obtained. Operation of calculating the depth of the sample from the surface of the object with respect to the light source may be performed simultaneously with or separately from operation of calculating the depth of the sample from the surface of the object with respect to the viewpoint.
  • the corresponding texel is detected from a texture image stored in the texture image storage unit 232.
  • the texture image can have brightness values and color values to which a translucency property and multi-layer tissue of the object have been reflected.
  • a brightness value to which a translucency property has been reflected can be obtained by calculating a depth from a surface of an object to a sample of which a depth has to be obtained with respect to a light source, and detecting the corresponding texel.
  • a color value to which multi-layer tissue has been reflected can be obtained by calculating a depth from a surface of an object to a sample of which a depth has to be obtained with respect to a viewpoint, and detecting the corresponding texel. In this way, the brightness and color values of the texel are obtained (560).
  • the obtained brightness and color values are accumulated in existing values, thereby performing rendering (570). Accordingly, the brightness and color values of a display pixel are decided by accumulating the brightness and color values of texels corresponding to samples sampled along a light beam passing through the corresponding display pixel. After the brightness and color values are accumulated, it is determined whether sampling needs to continue to be performed (580). If sampling needs to continue to be performed, the location of the next sample is acquired, and the above-described operation is performed on the next sample. If sampling does no longer need to continue to be performed, the process terminates.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Computer Graphics (AREA)
  • General Physics & Mathematics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Animal Behavior & Ethology (AREA)
  • Pathology (AREA)
  • Radiology & Medical Imaging (AREA)
  • Biomedical Technology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Surgery (AREA)
  • Biophysics (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Remote Sensing (AREA)
  • Radar, Positioning & Navigation (AREA)
  • General Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Computer Hardware Design (AREA)
  • Geometry (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Architecture (AREA)
  • Image Generation (AREA)

Description

  • Embodiments of the present invention relate to an ultrasonic imaging apparatus and a control method thereof.
  • An ultrasonic imaging apparatus emits ultrasonic signals towards a specific part of a subject (e.g., a human body) from the skin surface of the subject, and receives ultrasonic signals (that is, ultrasonic echo signals) reflected from the specific part of the subject so as to non-invasively acquire section images about soft tissue or images about blood vessels based on information about the ultrasonic echo signals.
  • Compared to other imaging apparatuses including an X-ray imaging apparatus, a Computerized Tomography (CT) scanner, a Magnetic Resonance Image (MRI) apparatus, and a radiographic diagnosis system, the ultrasonic imaging apparatus is a low-priced apparatus having a small size, can display images in real time, and have high safety since there is no risk for patients to be exposed to radiation such as X-rays. For the advantages, the ultrasonic imaging apparatus is widely used to diagnose the heart, breasts, abdomen, urinary organs, uterus, etc.
  • In general, an ultrasonic imaging apparatus includes a main body to accommodate main components of the ultrasonic imaging apparatus, an ultrasonic collector to transmit and receive ultrasonic waves, a control panel including various switches and keys through which commands for manipulating the ultrasonic imaging apparatus are input, and a display to display the results of ultrasonic diagnosis as images.
  • A method in which the ultrasonic imaging apparatus performs ultrasonography on a subject is as follows. First, an operator holds the ultrasonic collector with his/her one hand to move the ultrasonic collector while keeping the ultrasonic collector in contact with the skin surface of a subject, and manipulates the control panel with the other hand to perform ultrasonography. An ultrasonic image obtained by the ultrasonography is displayed in real time through the display so that the operator can diagnose the state of the subject.
  • However, since ultrasonic images obtained by ultrasonography are monochrome images due to the physical properties of the ultrasonic imaging apparatus, there is limitation in representing the inner structure of a subject realistically. In order to overcome the limitation, a method of producing 3-dimensional (3D) color ultrasonic images based on arbitrarily selected colors has been developed. However, the method still has a problem that reality of images is low since the images are not based on the real colors of a subject. Particularly, when producing 3D ultrasonic images of the fetus, the method results in lower reality than when producing images of the heart, breasts, etc.
  • US 2013/150719 relates to an ultrasound imaging system and method for ultrasound imaging. The method includes generating a volume-rendered image from three-dimensional ultrasound data. The volume-rendered image is colorized with at least two colors according to a depth-dependent color scheme. The method includes displaying the volume-rendered image. The method includes generating a planar image from the three-dimensional ultrasound data, where the planar image is colorized according to the same depth-dependent color scheme. The method includes displaying the planar image.
  • EP 2 441 389 relates to an ultrasonic diagnostic apparatus to generate a three-dimensional tomographic image including: an offset calculating section which increases or decreases the brightness value of each voxel according to the brightness value of each voxel of a three-dimensional tomographic image volume data. The amount of increase or decrease in the brightness value of each voxel of the offset calculating section is adjustable through a control panel, and a tomographic image volume rendering section generates the three-dimensional tomographic image on the basis of a three-dimensional tomographic image volume data in which the brightness value is offset by the offset calculating section.
  • The invention is defined by independent claims 1 and 7. Aspects or embodiments not falling under the scope of said claims do not form part of the present invention. Therefore, it is an aspect to provide an ultrasonic imaging apparatus for producing a realistic 3-dimensional (3D) image by performing rendering with reference to a texture image to which a translucency property or multi-layer tissue of a subject has been reflected, and a control method of the ultrasonic imaging apparatus.
  • Additional aspects will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the invention.
  • In accordance with one aspect, an ultrasonic imaging apparatus includes: an ultrasonic collector configured to collect ultrasonic waves from an object; a volume data generator configured to generate volume data based on the ultrasonic waves; and an image processor configured to perform volume rendering on the volume data with reference to a texture image, wherein a translucency property and multi-layer tissue of the object are reflected to each texel of the texture image.
  • In accordance with another aspect, an ultrasonic imaging apparatus includes: an ultrasonic collector configured to collect ultrasonic waves from an object; a volume data generator configured to generate volume data based on the ultrasonic waves; and an image processor configured to perform volume rendering on the volume data with reference to a texture image, wherein a translucency property of the object is reflected to each texel of the texture image.
  • In accordance with another aspect, an ultrasonic imaging apparatus includes: an ultrasonic collector configured to collect ultrasonic waves from an object; a volume data generator configured to generate volume data based on the ultrasonic waves; and an image processor configured to perform volume rendering on the volume data with reference to a texture image, wherein multi-layer tissue of the object is reflected to each texel of the texture image.
  • In accordance with another aspect, a control method of an ultrasonic imaging apparatus includes: collecting ultrasonic waves from an object; generating volume data based on the ultrasonic waves; and performing volume rendering on the volume data with reference to a texture image, wherein a translucency
    property and multi-layer tissue of the object are reflected to each texel of the texture image.
  • In accordance with another aspect, a control method of an ultrasonic imaging apparatus includes: collecting ultrasonic waves from an object; generating volume data based on the ultrasonic waves; and performing volume rendering on the volume data with reference to a texture image, wherein a translucency property of the object is reflected to each texel of the texture image.
  • In accordance with another aspect, a control method of an ultrasonic imaging apparatus includes: collecting ultrasonic waves from an object; generating volume data based on the ultrasonic waves; and performing volume rendering on the volume data with reference to a texture image, wherein multi-layer tissue of the object is reflected to each texel of the texture image.
  • Advantageous Effects of Invention
  • According to the ultrasonic imaging apparatus and the control method thereof as described above, the following effects can be obtained.
  • By rendering volume data after reflecting a translucency property and multi-layer tissue of a subject to the volume data, it is possible to control and produce a realistic image such that it looks close to the real structure of the subject.
  • Particularly, when the ultrasonic imaging apparatus is used in gynecological diagnosis, the ultrasonic imaging apparatus may display the fetus's figure realistically, which leads to improvement of emotional stability and mental satisfaction of fragment woman and her family.
  • Brief Description of Drawings
  • These and/or other aspects will become apparent and more readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:
    • FIG. 1 is a perspective view of an ultrasonic imaging apparatus according to an embodiment ;
    • FIG. 2 is a block diagram of a control configuration of an ultrasonic imaging apparatus according to an embodiment;
    • FIG. 3 is a block diagram of a control configuration of an image processor according to an embodiment;
    • FIG. 4 illustrates light paths when a light source is located to face a viewpoint with an object in between;
    • FIG. 5 illustrates an example of a texture image including information about a translucency property and multi-layer tissue of an object;
    • FIG. 6 is a view for describing a process of rearranging the position of an object with respect to a viewpoint upon sampling, according to an embodiment;
    • FIG. 7 is a view for describing a method of calculating the location of a pixel of a texture image, corresponding to each sample, according to an embodiment;
    • FIGS. 8A, 8B, and 8C are views for describing a method of calculating a depth from a surface to each sample with respect to a viewpoint, according to an embodiment;
    • FIG. 9 is a flowchart of a method of calculating a depth from an object's surface to each sample with respect to a viewpoint, according to an embodiment;
    • FIGS. 10A, 10B, and 10C are views for describing a method of calculating a depth from a surface to each sample with respect to a light source, according to an embodiment ;
    • FIG. 11 is a flowchart of a method of calculating a depth from an object's surface to each sample with respect to a light source, according to an embodiment; and
    • FIG. 12 is a flowchart of a rendering algorithm for a volume rendering method, according to an embodiment.
    Best Mode for Carrying out the Invention
  • Reference will now be made in detail to the embodiments, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to like elements throughout.
  • FIG. 1 is a perspective view of an ultrasonic imaging apparatus according to an embodiment. Referring to FIG. 1, the ultrasonic imaging apparatus may include a main body 100, an ultrasonic collector 110, an input unit 150, a main display unit 160, and a sub display unit 170.
  • In one side of the main body 100, one or more female connectors 145 may be provided. A male connector connected to a cable 130 may be physically coupled with one of the female connectors 144.
  • In the bottom of the main body 100, a plurality of castors (not shown) for moving the ultrasonic imaging apparatus may be provided. The plurality of castors may be used to fix the ultrasonic imaging apparatus at a specific location and to move it in a specific direction.
  • The ultrasonic collector 110 contacts the skin surface of an object, and may transmit or receive ultrasonic waves. More specifically, the ultrasonic collector 110 functions to irradiate transmission signals (that is, ultrasonic signals) received from the main body 100 to the inside of the object, receive ultrasonic echo signals reflected from a specific part of the object, and transfer the ultrasonic echo signals to the main body 100. The ultrasonic collector 110 may be connected to one end of the cable 130, and the other end of the cable 130 may be connected to the male connector 140. The male connector 140 connected to the other end of the cable 130 may be physically coupled with one of the female connectors 145 of the main body 145.
  • The input unit 150 is used to receive commands for instructing operations of the ultrasonic imaging apparatus. For example, a mode selection command, such as an Amplitude mode (A-mode), a Brightness mode (B-mode), or a Motion mode (M-mode), or an ultrasonography start command may be input through the input unit 150. A command input through the input unit may be transferred to the main body 100 through wired/wireless communication.
  • The input unit 150 may include at least one of a keyboard, a foot switch, and a foot pedal. The keyboard may be hardwarily implemented, and disposed on the upper part of the main body 100. The keyboard may include at least one(s) of a switch, keys, a joystick, and a trackball. As another example, the keyboard may be softwarily implemented as a graphic user interface (GUI). In this case, the keyboard may be displayed through the sub display unit 170 or the main display unit 160. The foot switch or the foot pedal may be provided in the lower part of the main body 100, and an operator may control operations of the ultrasonic imaging apparatus using the foot pedal.
  • An ultrasonic collector holder 120 for accommodating the ultrasonic collector 110 may be provided to be adjacent to the input unit 150. There may be provided one or more ultrasonic collector holders. The operator may put the ultrasonic collector 110 into the ultrasonic collector holder 120 to safely keep the ultrasonic collector 110 when he/she does not use the ultrasonic imaging apparatus.
  • The sub display unit 170 may be installed in the main body 100. FIG. 1 shows a case in which the sub display unit 170 is provided on the upper part of the input unit 150. The sub display unit 170 may display applications related to operations of the ultrasonic imaging apparatus. For example, the sub display unit 170 may display menus or guidance for ultrasonography. The sub display unit 170 may be implemented as a Cathode Ray Tube (CRT) or a Liquid Crystal Display (LCD).
  • The main display unit 160 may be also installed in the main body 100. FIG. 1 shows a case in which the main display unit 160 is positioned above the sub display unit 170. The main display unit 160 may display ultrasonic images acquired by ultrasonography. The main display 160 may be, like the sub display 170, implemented as a CRT or a LCD. In FIG. 1, the main display unit 160 is integrated with the main body 100, however, the main display unit 160 may be separated from the main body 100.
  • In the embodiment of FIG. 1, the ultrasonic imaging apparatus includes both the main display unit 160 and the sub display unit 170, however, the sub display unit 170 may be omitted. In this case, applications, menus, etc., which are displayed through the sub display unit 170, may be display through the main display unit 160.
  • FIG. 2 is a block diagram of a control configuration of an ultrasonic imaging apparatus according to an embodiment. Referring to FIGS. 1 and 2, the ultrasonic collector 110 includes a plurality of ultrasonic transducers 114 to generate ultrasonic waves according to alternating current applied from a power source 112, to irradiate the ultrasonic waves to an object, to receive echo ultrasonic waves reflected from a target part of the object, and to convert the received echo ultrasonic waves into electrical signals. The power source 112 may be an external power supply, or a power storage unit included in the ultrasonic imaging apparatus.
  • Each ultrasonic transducer 114 may be a magnetostrictive ultrasonic transducer using the magnetostrictive effect of a magnetic material, a piezoelectric ultrasonic transducer using the piezoelectric effect of a piezoelectric material, or a capacitive micromachined ultrasonic transducer (CMUT) that transmits and receives ultrasonic waves using vibration of several hundreds or thousands of micromachined thin films.
  • When alternating current from the power source 112 is applied to each ultrasonic transducer 114, the piezoelectric vibrator or thin film of the ultrasonic transducer 114 vibrates, thus generating ultrasonic waves. The generated ultrasonic waves are irradiated to an object, for example, to the inside of a human body. The irradiated ultrasonic waves are reflected by one or more target areas that may be located at different depths from the surface of the object. The ultrasonic transducers 114 receive ultrasonic echo signals reflected from the target areas, and convert the received ultrasonic echo signals into electrical signals to obtain a plurality of received signals.
  • The received signals are transferred to the main body 100 through a wired/wireless communication network. Since the ultrasonic collector 110 receives the ultrasonic echo signals through a plurality of channels, the plurality of received signals are also transferred to the main body 100 through a plurality of channels.
  • The main body 100 may include a beamformer 210, a volume data generator 220, and an image processor 230. In another embodiment, one or more components of the above-mentioned components may be omitted.
  • The beamformer 210 may perform beamforming based on the plurality of received signals. Beamforming is operation of focusing a plurality of received signals received through a plurality of channels to acquire a proper ultrasonic image about the inside of an object.
  • The beamformer 210 corrects time differences of a plurality of received signals, caused by distance differences between the respective transducers 114 and a target area of the object. Then, the beamformer 210 enhances a plurality of received signals corresponding to specific channels, or relatively attenuates a plurality of received signals corresponding to other channels, thereby focusing the plurality of received signals. At this time, the beamformer 210 may enhance or attenuate a specific received signal by selectively allocating predetermined weights to a plurality of received signals received through individual channels.
  • The beamformer 210 may focus a plurality of received signals collected by the ultrasonic collector 110, for each frame, in consideration of the locations and focusing points of the transducers 114 of the ultrasonic collector 110.
  • The beamforming that is performed by the beamformer 210 may be one of data-independent beamforming and adaptive beamforming.
  • The volume data generator 220 generates ultrasonic image data based on the signals focused by the beamformer 210. If the beamformer 210 focuses a plurality of received signals for each of a plurality of frames, the volume data generator 220 generates a plurality of pieces of ultrasonic image data respectively corresponding to the respective frames based on a plurality of focused signals. The ultrasonic image data may be 3D ultrasonic image data for the object, that is, volume data for the object.
  • The image processor 230 may receive the volume data for the object from the volume data generator 230. Then, the image processor 230 performs volume rendering on the volume data, and transfers the results of the volume rendering to the main display unit 160 so that the main display unit 160 can output the results of the volume rendering as images.
  • The image processor 230 may volume-render the 3D volume data using one of volume rendering methods well-known in the art. In detail, volume rendering may be classified into surface rendering and direct volume rendering.
  • The surface rendering is to extract surface information from volume data based on predetermined scalar values and amounts of spatial changes, to convert the surface information into a geometric factor, such as a polygon or a surface patch, and then to apply a conventional rendering technique to the geometric factor. Examples of the surface rendering are a marching cubes algorithm and a dividing cubes algorithm.
  • The direct volume rendering is to directly render volume data without converting volume data into a geometric factor. The direct volume rendering is useful to represent a translucent structure since it can visualize the inside of an object as it is. The direct volume rendering may be classified into an object-order method and an image-order method according to a way of approaching volume data.
  • The image-order method is to sequentially decide pixel values of an image. An example of the image-order method is volume ray casting. According to the volume ray casting, as illustrated in FIG. 6, a virtual ray is irradiated from a specific viewpoint toward a predetermined pixel of a display screen, and then voxels through which the virtual ray has been transmitted from among voxels of volume data are detected. Then, brightness values of the detected voxels are accumulated to decide a brightness value of the corresponding pixel of the display screen. Alternatively, an average value of the detected voxels may be decided as a brightness value of the corresponding pixel of the display screen. Also, a weighted average value of the detected voxels may be decided as a brightness value of the corresponding pixel of the display screen.
  • The object-order method is to directly project volume data on an image. An example of the object-order method is splatting. According to splatting, by aligning the center of a reconfigured kernel with a point at which a voxel is projected on a plane to be observed, and then causing peripheral pixels to distribute values to the voxel, a brightness value of the corresponding pixel is decided.
  • If the volume rendering is completed, a 3D ultrasonic image may be produced. Each pixel of the 3D ultrasonic image may have a brightness value and a color value. The following description is given under an assumption that the volume ray casting of the above-mentioned volume rendering methods is used.
  • Referring again to FIG. 2, the storage unit 240 may store data or algorithms required for operations of the ultrasonic imaging apparatus 200. For example, the storage unit 240 may store an algorithm for rendering volume data, an algorithm for calculating the location of a sample with respect to a viewpoint, and an algorithm for calculating a depth from a surface to each sample.
  • The storage unit 240 may be Read Only Memory (ROM), Random Access Memory (RAM), Programmable Read Only Memory (PROM), Erasable Programmable Read Only Memory (EPROM), a non-volatile memory device such as flash memory, a volatile memory device such as Random Access Memory (RAM), or an optical disk. However, the storage unit 280 is not limited to these, and may be any other storage device well-known in the art.
  • An embodiment in which the main body 100 of the ultrasonic imaging apparatus includes the beamformer 210, the volume data generator 220, the image processor 230, and the storage unit 240 has been described above with reference to FIGS. 1 and 2. However, the main body 100 does not need to include all the above-mentioned components. In other words, some of the components may be included in another component, for example, in the ultrasonic collector 110.
  • For example, the ultrasonic collector 110 may include the beamformer 210 to produce an ultrasonic image based on electrical signals, and to correct the produced ultrasonic image.
  • The above description relates to the appearance and control configuration of an ultrasonic imaging apparatus according to an embodiment.
  • Hereinafter, a control configuration of the image processor 230 of the ultrasonic imaging apparatus, and the functions of the individual components of the ultrasonic imaging apparatus will be described in more detail with reference to FIGS. 3 to 12.
  • FIG. 3 is a block diagram of a control configuration of the image processor 230 according to an embodiment. Referring to FIG. 3, the image processor 230 may include a texture image storage unit 232, a sampling unit 234, a calculator 236, and a data accumulating unit 238.
  • The texture image storage unit 232 may store texture images that are referred to upon volume rendering. The texture image storage unit 232 may be implemented as the same type of device as the storage unit 240 since the texture image storage unit 232 also functions to store data.
  • The texture image storage unit 232 may store one or more texture images. A texture image that is to be stored in the texture image storage unit 232 may reflect a translucency property or multi-layer tissue of an object. Hereinafter, a translucency property or multi-layer tissue of an object will be described.
  • A phenomenon in which light is transmitted through a certain object without being absorbed in or reflected from the object is called transmission. A degree of light transmission greatly depends on the properties of an object through which the light is transmitted. For example, a transparent material such as glass transmits a major part of incident light although it reflects a small amount of the incident light and absorbs a small amount of the incident light. A degree at which light is transmitted through a material is determined as transparency of the material, and a material through which light is 100% transmitted can be considered to have transparency. Meanwhile, a material through which no light is transmitted is considered as an opaque material. If a certain material transmits a small amount of light, and scatters light according to its surface structure and properties to lower transparency, the material is considered to have a translucency property.
  • An object such as the fetus having soft tissue can be considered to have a translucency property since a constant amount of light can be transmitted through the object. Accordingly, such an object is strongly dependent on an effect due to a translucency property since a part of light irradiated from a light source is transmitted deeply to tissue. In the case of an object having a translucency property, an effect due to light transmission through the object must be considered even when a light source is positioned behind the object, as well as when a light source is positioned in front of the object. A translucency property can be represented through calculations under an assumption that a part of incident light is transmitted through the inside of an object while the remaining part of the incident light is reflected from the surface of the object.
  • FIG. 4 illustrates light paths when a light source is located to face a viewpoint with an object in between. In FIG. 4, no path of light reflected from the object is shown since the reflected light has no influence in observing the object at the viewpoint. First, a path of light passing a point df1 at which light from a light source enters the inside of the object, and a point do1 at which the light transmitted through the inside of the object is escaped from the object is described below. The light is transmitted through the inside of the object by a distance of S1. Meanwhile, when light generated from a light source is transmitted through an object having a non-uniform thickness, a distance by which the light is transmitted through the inside of the object depends on a path of the light.
  • Referring to FIG. 4, a distance S1 by which light is transmitted through the lower part of the object is longer than a distance S2 by which light is transmitted through the upper part of the object (S1>S2). A part of light incident to an object is absorbed in the object when the light is transmitted through the inside of the object, so long as the object is not a completely transparent material. Accordingly, as a distance by which light is transmitted through the inside of an object is longer, the more amount of light is absorbed in the object. Therefore, in the case of FIG. 4, when the object is seen from the viewpoint, the point do1 is observed darker than the point do2. That is, an area having the deeper depth from the surface of an object which a light source faces must be represented darker than an area having the shallower depth from the surface of the object which the light source faces in order to improve the reality of the object. In consideration of this, one axis of a texture image that is to be stored in the texture image storage unit 232 may be configured to represent a depth from a surface of an object which a light source faces.
  • With the translucency property of an object, multi-layer tissue of the object must be considered. Light transmitted through a medium is influenced by the kind of the medium. Specifically, since different kinds of mediums absorb different wavelengths of light, the color of an object shown from a viewpoint depends on a depth to which light has been transmitted through the object. Accordingly, when an object is shown from a viewpoint, the effect of light transmitted through different material layers of the object, that is, the colors of light transmitted through the different material layers of the object are mixed and visually shown.
  • For example, if an object is a human body, multi-layer tissue of the human body can be considered. The reason is because the skin of a human body is not a single material and consists of several different kinds of layers. If the skin of the human body is divided into a horny layer, an epidermis layer, a dermis layer, and a panniculus layer, the object may be represented in a different way from an object having a single layer since the respective layers have different properties. Accordingly, when an object has multi-lay tissue like the skin of a human body, the multi-layer tissue of the object is considered upon rendering in order to improve reality. Particularly, in the case of volume rendering, since the inside of an object as well as the surface of the object must be rendered, a method of deciding color values in consideration of multi-layer tissue of an object can improve the reality of rendering.
  • In order to consider multi-layer tissue of an object, what layer of the object a sample is located in needs to be determined. In order to determine a layer to which a sample belongs, a depth at which a sample is located from a surface with respect to a viewpoint may be calculated. By deciding a color value in consideration of the depth of a sample upon rendering, reality can be improved. In order to reflect the depths of samples to a texture image, one axis of the texture image may be configured to represent a depth from a surface with respect to a viewpoint.
  • FIG. 5 illustrates an example of a texture image including information about a translucency property and multi-layer tissue of an object (e.g., the fetus). In the following description, a texture image, which is shown as black and white in FIG. 5, is assumed to be in color since the present disclosure is implemented using a color texture image.
  • The texture image storage unit 232 (see FIG. 3) may store 2-dimensional (2D) texture images. Each of the 2D texture images may be configured with a group of texels each having a brightness value or a color value. When rendering is performed, the brightness or color values of the texels may be referred to in order to improve reality. The texel is a minimum unit constructing text image data.
  • Referring to FIG. 5, a texture image may be configured such that the horizontal axis represents the depths of samples from a surface with respect to a viewpoint, and the vertical axis represents the depths of the samples from a surface with respect to a light source. As described above, the depth (the horizontal axis of a texture image) of a sample from a surface with respect to a viewpoint is used to determine what layer of the corresponding object the sample is located in, and accordingly, the horizontal axis of the texture image may become criteria to decide a color value of a texel. Also, the depth (the vertical axis of a texture image) of a sample from a surface with respect to a light source is used to determine a depth to which light has been transmitted through the corresponding object, and accordingly, the vertical axis of the texture image may become criteria to decide a brightness value of a texel.
  • Referring again to FIG. 3, the sampling unit 234 receives volume data of an object from the volume data generator 220 (see FIG. 2). The sampling unit 234 performs sampling for volume rendering on the received volume data.
  • Upon sampling, the position of the object may be rearranged with respect to a viewpoint in order to facilitate calculations for volume rendering. Specifically, axes of a display screen may be set to axes of an orthogonal coordinates system, and the locations of samples may be represented as vectors. Thereby, distances between samples and a surface may be obtained through a simple mathematical calculation.
  • FIG. 6 is a view for describing a process of rearranging the position of an object with respect to a viewpoint upon sampling, according to an embodiment. FIG. 6 shows an example of an arrangement of volume data samples, having a cube shape in which each side has scalar values of N voxels. Each voxel is represented in the form of i =(i,j,k), and a voxel value of the vector i is F( i ) , wherein f is a function representing a voxel value corresponding to a voxel. First, it is assumed that light beams are projected from an observer toward volume data. In this case, a display screen is configured with MxN pixels, wherein M is the number of pixels arranged on the horizontal axis of the display screen, and N is the number of pixels arranged on the vertical axis of the display screen. Each light beam is projected to a pixel. Each pixel or a light beam projected through each pixel is represented as a vector u = (u,v), wherein u and v are coordinates on x- and y-axis, and points that are sampled on the light beams are represented as U = (u,v,w), wherein w is a coordinate on the z-axe. The z-axis coordinate w of the vector U increases in proportion to a sampling distance and the number of times of sampling. By obtaining the vector U , the locations of samples can be acquired. By representing the locations of samples as vectors, distances of the samples to a surface can be easily acquired by a mathematical calculation.
  • Setting values for sampling that is performed by the sampling unit 234 may be input by a user through the input unit 150. For example, a sampling distance, the number of times of sampling, and a point at which sampling terminates may be set to desired values by a user.
  • Referring again to FIG. 3, the calculator 236 receives the location of each sample, acquired by sampling, from the sampling unit 234. The calculator 236 functions to calculate a distance of the sample to the surface, using the received location of the sample. In order to increase the speed and accuracy of calculation, the calculator 236 may be hardwarily implemented to enable parallel processing using a plurality of pipelines.
  • The calculator 236 may perform a calculation for detecting a texel corresponding to each sample in the direction of the light beam passing through the display screen. Referring to FIG. 6, the horizontal axis of the texture image may be set to a depth from a surface with respect to a viewpoint, and the vertical axis of the texture image may be set to a depth from a surface with respect to a light source. By calculating a value of the sample on the horizontal or vertical axis of the texture image using the sample location received from the sampling unit 234, the corresponding texel can be detected. The calculator 236 may transfer a brightness value or a color value of the detected texel to the data accumulating unit 238.
  • FIG. 7 is a view for describing a method of calculating the location of a texel corresponding to each sample, according to an embodiment. In order to obtain a color value of each sample of an object, to which multi-layer tissue of the object has been reflected, a depth X of the corresponding sample from a surface with respect to a viewpoint is calculated. Also, in order to obtain a brightness value of each sample of an object, to which the translucency property of the object has been reflected, a depth Y of the corresponding sample from a surface with respect to a light source is calculated. Based on the resultant coordinate value (x, y), a texel corresponding to the sample may be detected from the texture image of FIG. 6. The calculator 236 may transfer the color value and brightness value of the detected texel to the data accumulating unit 238. Hereinafter, a method of calculating the depths X and Y will be described in detail with reference to FIGS. 8 to 11.
  • FIGS. 8A, 8B, and 8C are views for describing a method of calculating a depth X from a surface to each sample with respect to a viewpoint, according to an embodiment. Referring to FIG. 8A, when a virtual light beam is projected to an object from a viewpoint, volume data is sampled in the direction of the light beam. In order to find a surface on the path of the light beam, sampling from the viewpoint continues to be performed, and if it is found a first sample which is not an empty space, the location of the found sample is estimated as the location of a surface. If an empty space is found during continuous sampling, the location of a sample which is not an empty space and is first found is again estimated as the location of a surface.
  • Referring to FIG. 8B, a gradient vector of the estimated surface is calculated. The gradient vector of the surface is calculated using texel values of peripheral samples on x-, y-, and z-axis. The following Equation (1) is used to calculate an x component of a gradient vector of a surface. In Equation (1), a vector represents the location of a sample, f represents a voxel value of the sample, and h represents a unit distance between samples, wherein u, v, and w are x-, y-, and z-axis coordinates. g x U u v w = f U u + 1 , v , w f U u 1 , v , w 2 h
    Figure imgb0001
  • Referring to FIG. 8C, a plane in which the normal vector is the gradient vector of the surface is obtained. The plane is a tangential plane of the object at a point corresponding to the surface. A distance from the plane to the sample of which a depth has to be calculated is calculated, and the distance becomes a depth X from the surface to the sample with respect to the viewpoint.
  • FIG. 9 is a flowchart of a method of calculating a depth X from an object's surface to each sample with respect to a viewpoint, according to an embodiment. The depth X can be calculated by obtaining a tangential plane at a point corresponding to the surface, and calculating a distance from the tangential plane to the sample.
  • First, sampling is performed along a light beam from a viewpoint (300). At this time, the number of times of sampling increases in proportion to a degree of reality of volume rendering. The reason is because as the number of times of sampling is more, the more accurate location of a surface can be found. The number of times of sampling may be set to a value input by a user through the input unit 150 (see FIG. 2). Sampling may be performed simultaneously with or separately from the following process, and in the following description, it is assumed that sampling is performed simultaneously with the following process.
  • Successively, voxel values of samples are checked in order from the viewpoint (310). Then, it is determined whether each sample is an empty space based on the voxel values (320). Whether each sample is an empty space may be determined using an Opacity Transfer Function (OTF) that defines the relationship between density values and transparency among the voxel values. Since a method of determining an empty space using an OTF is well-known in the art, further descriptions will be omitted.
  • If a sample is an empty space, the next sample is checked (325). If a sample which is not an empty space is found, the location values of the sample are acquired (330). The sample whose location values have been acquired is estimated as a surface. The reason is because a sample which is not an empty space and is first found is a first sample that exists in the inside of the object in the direction of the light beam.
  • Then, operation for calculating a depth X of the sample is performed. The depth X of the sample is calculated using the location values of the sample. To do this, operation of calculating a gradient vector at the location of the sample is performed (340). The gradient vector is calculated using peripheral values on the x-, y-, and z-axis of the sample. Since a method of calculating the gradient vector has been described above with reference to FIG. 8B, further descriptions will be omitted.
  • After the gradient value is calculated, a plane on which the location of the sample estimated as the surface exists and in which the normal vector is the gradient vector of the surface is obtained (350). The plane is a tangential plane of the object at a point corresponding to the surface since the plane meets the object only at the point of the surface and is perpendicular to the object. By using the tangential plane which is perpendicular to the object, the depths of samples can be easily calculated in the direction of the light beam.
  • Finally, by calculating a distance between the tangential plane and the sample of which the depth X has to be calculated, the depth X of the sample from the surface with respect to the viewpoint may be acquired (360). The depth X may be used to detect a texel on the horizontal axis of the texture image of FIG. 6.
  • FIGS. 10A, 10B, and 10C are views for describing a method of calculating a depth Y from a surface to each sample with respect to a light source, according to an embodiment. Referring to FIG. 10A, a light beam advancing from a sample of which a depth has to be calculated toward a light source is generated. This is aimed at calculating a distance by which light irradiated from a light source is transmitted through a medium until the light arrives at the corresponding sample to reflect the translucency property of the corresponding object. Since light reflected from the surface of the object has no influence in deciding a brightness value of the sample, only light advancing straightly is considered.
  • Referring to FIG. 10B, volume data is sampled at regular time intervals along the generated light beam. At this time, sampling is performed along the traveling direction of the light beam, starting from the sample of which the depth has to be calculated, so that brightness values among voxel values of samples are checked. Since a brightness value becomes greater as light irradiated from a light source is transmitted through a medium to the deeper depth, the brightness values are checked to detect a sample on the surface of a medium. The brightness values acquired in order of sampling are compared to a predetermined threshold value. Since a point having a great change in brightness values can be considered as a point at which light starts to be transmitted through a medium, if the brightness value of a certain sample falls below the predetermined threshold value, the previous sample may be estimated as a surface. Then, a distance between the surface and the sample of which the depth has to be calculated is calculated, so that a depth Y from the surface to the sample can be calculated.
  • FIG. 11 is a flowchart of a method of calculating a depth Y from an object's surface to each sample with respect to a light source, according to an embodiment. The depth Y can be calculated based on a fact that the strength of light varies depending on a distance by which the light is transmitted through a medium until the light arrives at a sample.
  • In order to calculate the depth Y, first, a light beam advancing from a sample of which a depth has to be calculated toward a light source is generated (400). This is aimed at setting a path of light arriving at the corresponding sample from an actual light source, and easily calculating a distance by which the light advancing along the path is transmitted through a medium.
  • After the light beam is generated, sampling is performed at regular time intervals along the traveling direction of the generated light beam (410). At this time, the number of times of sampling increases in proportion to a degree of reality of volume rendering. The reason is because as the number of times of sampling is more, the more accurate location of a surface can be found. The number of times of sampling may be set to a value input by a user through the input unit 150 (see FIG. 2). Sampling may be performed simultaneously with or separately from the following process, and in the following description, it is assumed that sampling is performed simultaneously with the following process.
  • If sampling is performed, brightness values among voxel values of samples are acquired in order, starting from a sample located most distant from the light source (420), and the brightness values are compared to a predetermined threshold value (430). By comparing the brightness values to the predetermined threshold value, it can be determined whether a sample is included in the inside of the object. The predetermined threshold value may be set to an appropriate value within an error range of a brightness value when light is not transmitted through any medium.
  • When the brightness value of a sample is compared to the predetermined threshold value, if the brightness value of the sample is greater than the predetermined threshold value, the corresponding sample is determined as a sample located in the inside of the object, and then, the next sample is compared to the predetermined threshold value (435). Meanwhile, if the brightness value of the sample is smaller than the predetermined threshold value, the sample is determined as a sample located outside the object.
  • If a sample is determined as a sample located outside the object, the previous sample is estimated as a surface of the object. Then, the location values of the sample are acquired (440). The reason is because if a sample located outside the object is found after successive samples located in the inside of the object are found, the previous sample of the sample located outside the object is a final sample located in the inside of the object in the traveling direction of the light beam.
  • Then, a distance from the surface to the sample of which the depth has to be calculated is calculated, so that a depth Y from the surface to the sample can be calculated (450). The depth Y may be used to detect a texel on the vertical axis of the texture image of FIG. 6.
  • Referring to FIGS. 3 and 11, the calculator 236 may detect a texel corresponding to the sample from a texture image stored in the texture image storage unit 232 based on the resultant depth value (x, y). Then, a brightness value or a color value is detected from the corresponding texel.
  • The data accumulating unit 238 receives the brightness value or the color value of the texel detected by the calculator 236. The brightness values or color values of texels corresponding to samples acquired along the path of a light beam passing through a display pixel with respect to a viewpoint are all accumulated in the data accumulating unit 238. Based on the values accumulated in the accumulating unit 238, the brightness or color of the display pixel through which the light beam passes is decided.
  • FIG. 12 is a flowchart of a rendering algorithm for a volume rendering method, according to an embodiment.
  • First, sampling is performed at regular time intervals in the direction of a light beam from a viewpoint (500). At this time, as sampling is performed at shorter time intervals, more samples are subject to a calculation to decide a brightness value and a color value of each display pixel. Accordingly, since more brightness values and color values of texels are acquired, a more amount of data is accumulated to obtain the results of rendering having high reality. A sampling interval may be set to a value input by a user through the input unit 150.
  • Then, voxel values of samples are checked in order, starting from a sample located closest to a viewpoint (510). Then, it is determined whether each sample is an empty space (520). Whether the sample is an empty space may be determined using the OTF which is well-known in the art. If it is determined that a sample is an empty space based on the voxel value of the sample, it is determined whether the next sample is an empty space (525). The reason is because a sample corresponding to an empty space has no information about an object.
  • If a sample which is not an empty space is found, the location values of the sample are acquired (530), and a calculation for detecting a texel corresponding to the sample from a texture image is performed. To do the calculation, a depth of the sample from a surface of the object with respect to a light source is calculated (540), and a depth of the sample from a surface of the object with respect to the viewpoint is calculated (550).
  • The depth of the sample from the surface of the object with respect to the light source is calculated by calculating a distance by which a light beam generated from the sample passes through a medium until the light beam arrives at the light source. The depth of the sample from the surface of the object with respect to the viewpoint is calculated by estimating a sample which is not an empty space and is first found as a surface, and calculating a distance between a tangential plane of the object at a point corresponding to the surface and the sample of which the depth has to be obtained. Operation of calculating the depth of the sample from the surface of the object with respect to the light source may be performed simultaneously with or separately from operation of calculating the depth of the sample from the surface of the object with respect to the viewpoint.
  • Based on the results of the calculations, the corresponding texel is detected from a texture image stored in the texture image storage unit 232. The texture image can have brightness values and color values to which a translucency property and multi-layer tissue of the object have been reflected. A brightness value to which a translucency property has been reflected can be obtained by calculating a depth from a surface of an object to a sample of which a depth has to be obtained with respect to a light source, and detecting the corresponding texel. A color value to which multi-layer tissue has been reflected can be obtained by calculating a depth from a surface of an object to a sample of which a depth has to be obtained with respect to a viewpoint, and detecting the corresponding texel. In this way, the brightness and color values of the texel are obtained (560).
  • The obtained brightness and color values are accumulated in existing values, thereby performing rendering (570). Accordingly, the brightness and color values of a display pixel are decided by accumulating the brightness and color values of texels corresponding to samples sampled along a light beam passing through the corresponding display pixel. After the brightness and color values are accumulated, it is determined whether sampling needs to continue to be performed (580). If sampling needs to continue to be performed, the location of the next sample is acquired, and the above-described operation is performed on the next sample. If sampling does no longer need to continue to be performed, the process terminates.

Claims (12)

  1. An ultrasonic imaging apparatus comprising:
    an ultrasonic collector (110) configured to collect ultrasonic waves from an object;
    a volume data generator (220) configured to generate volume data based on the ultrasonic waves; and
    an image processor (230) configured to perform volume rendering on the volume data with reference to a texture image,
    wherein a translucency property of the object is reflected to each texel of the texture image,
    characterised in that each texel to which the translucency property of the object has been reflected has a brightness value varying depending on a depth from a surface of the object with respect to a light source.
  2. The ultrasonic imaging apparatus according to claim 1, wherein the image processor (230) comprises a calculator (236) configured to detect a texel corresponding to a sample obtained by sampling the volume data, from the texture image.
  3. The ultrasonic imaging apparatus according to claim 2, wherein the calculator (236) is configured to detect the texel by calculating a distance from the surface of the object to the sample with respect to the light source.
  4. The ultrasonic imaging apparatus according to any one of the previous claims, wherein multi-layer tissue of the object is further reflected to each texel of the texture image.
  5. The ultrasonic imaging apparatus according to claim 4, wherein each texel to which the multi-layer tissue of the object has been reflected has a color value varying depending on a depth from a surface of the object with respect to a viewpoint.
  6. The ultrasonic imaging apparatus according to claim 5, wherein the calculator (236) is configured to detect the texel by calculating a distance from the surface of the object to the sample with respect to the viewpoint.
  7. A control method of an ultrasonic imaging apparatus, comprising:
    collecting ultrasonic waves from an object;
    generating volume data based on the ultrasonic waves; and
    performing volume rendering on the volume data with reference to a texture image,
    wherein a translucency property of the object is reflected to each texel of the texture image,
    characterised in that each texel to which the translucency property of the object has been reflected has a brightness value varying depending on a depth from a surface of the object with respect to a light source
  8. The control method according to claim 7, wherein the performing of the volume rendering comprises detecting a texel corresponding to a sample obtained by sampling the volume data, from the texture image.
  9. The control method according to claim 8, wherein the detecting of the texel comprises detecting the texel by calculating a distance from the surface of the object to the sample with respect to the light source.
  10. The control method according to claims 7 to 9, wherein multi-layer tissue of the object is further reflected to each texel of the texture image.
  11. The control method according to claim 10, wherein each texel to which the multi-layer tissue of the object has been reflected has a color value varying depending on a depth from a surface of the object with respect to a viewpoint.
  12. The control method according to claim 11, wherein the detecting of the texel comprises detecting the texel by calculating a distance from the surface of the object to the sample with respect to a viewpoint.
EP14819625.6A 2013-07-05 2014-07-04 Ultrasonic imaging apparatus and control method thereof Active EP3017428B1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1020130078794A KR101851221B1 (en) 2013-07-05 2013-07-05 ultrasonic imaging apparatus and method for controlling a ultrasonic imaging apparatus using the same
PCT/KR2014/006007 WO2015002498A1 (en) 2013-07-05 2014-07-04 Ultrasonic imaging apparatus and control method thereof

Publications (3)

Publication Number Publication Date
EP3017428A1 EP3017428A1 (en) 2016-05-11
EP3017428A4 EP3017428A4 (en) 2017-02-22
EP3017428B1 true EP3017428B1 (en) 2020-09-02

Family

ID=52144009

Family Applications (1)

Application Number Title Priority Date Filing Date
EP14819625.6A Active EP3017428B1 (en) 2013-07-05 2014-07-04 Ultrasonic imaging apparatus and control method thereof

Country Status (5)

Country Link
US (1) US10535184B2 (en)
EP (1) EP3017428B1 (en)
KR (1) KR101851221B1 (en)
CN (1) CN105359191B (en)
WO (1) WO2015002498A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11100138B2 (en) * 2015-09-11 2021-08-24 Ayasdi Ai Llc Network representation for evolution of clusters and groups
KR102651126B1 (en) 2016-11-28 2024-03-26 삼성전자주식회사 Graphic processing apparatus and method for processing texture in graphics pipeline

Family Cites Families (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7194117B2 (en) 1999-06-29 2007-03-20 The Research Foundation Of State University Of New York System and method for performing a three-dimensional virtual examination of objects, such as internal organs
US6331116B1 (en) 1996-09-16 2001-12-18 The Research Foundation Of State University Of New York System and method for performing a three-dimensional virtual segmentation and examination
US6343936B1 (en) 1996-09-16 2002-02-05 The Research Foundation Of State University Of New York System and method for performing a three-dimensional virtual examination, navigation and visualization
US6450962B1 (en) 2001-09-18 2002-09-17 Kretztechnik Ag Ultrasonic diagnostic methods and apparatus for generating images from multiple 2D slices
US6825838B2 (en) 2002-10-11 2004-11-30 Sonocine, Inc. 3D modeling system
US7426734B2 (en) 2003-10-24 2008-09-16 Microsoft Corporation Facilitating presentation functionality through a programming interface media namespace
EP1542167A1 (en) * 2003-12-09 2005-06-15 Koninklijke Philips Electronics N.V. Computer graphics processor and method for rendering 3D scenes on a 3D image display screen
US7460117B2 (en) * 2004-05-25 2008-12-02 Siemens Medical Solutions Usa, Inc. Sliding texture volume rendering
JP4847334B2 (en) 2004-09-13 2011-12-28 株式会社日立メディコ Ultrasonic imaging apparatus and projection image generation method
US7764818B2 (en) 2005-06-20 2010-07-27 Siemens Medical Solutions Usa, Inc. Surface parameter adaptive ultrasound image processing
KR100738107B1 (en) 2006-02-15 2007-07-12 삼성전자주식회사 Apparatus and method for modeling based on 3 dimension point
US8049752B2 (en) * 2006-07-25 2011-11-01 Siemens Medical Solutions Usa, Inc. Systems and methods of determining sampling rates for volume rendering
JP4588736B2 (en) 2007-04-12 2010-12-01 富士フイルム株式会社 Image processing method, apparatus, and program
US7940268B2 (en) * 2007-06-29 2011-05-10 Microsoft Corporation Real-time rendering of light-scattering media
US7912264B2 (en) 2007-08-03 2011-03-22 Siemens Medical Solutions Usa, Inc. Multi-volume rendering of single mode data in medical diagnostic imaging
JP5253893B2 (en) 2008-06-03 2013-07-31 株式会社東芝 Medical image processing apparatus, ultrasonic diagnostic apparatus, and ultrasonic image acquisition program
US8224053B2 (en) 2009-03-31 2012-07-17 General Electric Company Methods and systems for displaying quantitative segmental data in 4D rendering
JP5730196B2 (en) * 2009-06-10 2015-06-03 株式会社日立メディコ Ultrasonic diagnostic apparatus, ultrasonic image processing apparatus, and ultrasonic image generation method
WO2011087035A1 (en) 2010-01-15 2011-07-21 株式会社 日立メディコ Ultrasonic diagnostic device and ultrasonic image display method
US20120095341A1 (en) * 2010-10-19 2012-04-19 Toshiba Medical Systems Corporation Ultrasonic image processing apparatus and ultrasonic image processing method
CN103338707B (en) 2011-01-26 2015-09-30 株式会社日立医疗器械 Diagnostic ultrasound equipment and image processing method
JP5818531B2 (en) 2011-06-22 2015-11-18 株式会社東芝 Image processing system, apparatus and method
US20130150719A1 (en) * 2011-12-08 2013-06-13 General Electric Company Ultrasound imaging system and method
JP5981281B2 (en) * 2012-09-12 2016-08-31 東芝メディカルシステムズ株式会社 Ultrasonic diagnostic apparatus, image processing apparatus, and program
CN104812312B (en) * 2012-09-26 2016-11-16 株式会社日立制作所 Diagnostic ultrasound equipment and ultrasound wave 3-D view creation method
WO2014168249A1 (en) * 2013-04-12 2014-10-16 日立アロカメディカル株式会社 Ultrasonic diagnostic device and ultrasonic three-dimensional image generation method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
None *

Also Published As

Publication number Publication date
KR101851221B1 (en) 2018-04-25
CN105359191A (en) 2016-02-24
US10535184B2 (en) 2020-01-14
KR20150005799A (en) 2015-01-15
CN105359191B (en) 2018-11-06
EP3017428A4 (en) 2017-02-22
WO2015002498A1 (en) 2015-01-08
US20170249777A1 (en) 2017-08-31
EP3017428A1 (en) 2016-05-11

Similar Documents

Publication Publication Date Title
CN102300505B (en) Ultrasonic diagnostic device and control program for displaying image data
JP6147489B2 (en) Ultrasonic imaging system
JP5265850B2 (en) User interactive method for indicating a region of interest
KR102111626B1 (en) Image processing apparatus and image processing method
KR102539901B1 (en) Methods and system for shading a two-dimensional ultrasound image
WO2007043310A1 (en) Image displaying method and medical image diagnostic system
CN111093518B (en) Ultrasound system for extracting image planes from volumetric data using touch interactions with images
US20180064420A1 (en) Apparatus for processing ultrasonic image and method thereof
US20150289837A1 (en) Ultrasonic apparatus and control method for the same
US10282893B2 (en) Imaging apparatus and imaging method
US9589387B2 (en) Image processing apparatus and image processing method
EP1952359A1 (en) System and method for generating for display two-dimensional echocardiography views from a three-dimensional image
US9449425B2 (en) Apparatus and method for generating medical image
KR20140131808A (en) Ultrasound imaging apparatus and control method for the same
EP3017428B1 (en) Ultrasonic imaging apparatus and control method thereof
JP6890677B2 (en) A virtual light source embedded in a 3D volume and coupled to the crosshairs in the MPR diagram
CN115444452A (en) Three-dimensional shear wave elastic imaging method and ultrasonic imaging system
EP3843637B1 (en) Ultrasound system and methods for smart shear wave elastography
CN112689478B (en) Ultrasonic image acquisition method, system and computer storage medium
JP7062384B2 (en) Medical image processing equipment and medical image processing programs
JP2000300555A (en) Ultrasonic image processing device
BE Surface Reconstruction in 3D Medical Imaging
KR20170067444A (en) Ultrasound diagnostic apparatus and control method for the same

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20151214

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

DAX Request for extension of the european patent (deleted)
A4 Supplementary search report drawn up and despatched

Effective date: 20170120

RIC1 Information provided on ipc code assigned before grant

Ipc: G06T 15/08 20110101ALI20170116BHEP

Ipc: G06T 19/00 20110101ALI20170116BHEP

Ipc: A61B 8/08 20060101ALI20170116BHEP

Ipc: G01S 15/89 20060101ALI20170116BHEP

Ipc: A61B 8/00 20060101AFI20170116BHEP

REG Reference to a national code

Ref country code: DE

Ref legal event code: R079

Ref document number: 602014069774

Country of ref document: DE

Free format text: PREVIOUS MAIN CLASS: G06T0017100000

Ipc: A61B0008000000

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: GRANT OF PATENT IS INTENDED

RIC1 Information provided on ipc code assigned before grant

Ipc: A61B 8/00 20060101AFI20200205BHEP

Ipc: G01S 15/89 20060101ALI20200205BHEP

Ipc: G06T 15/08 20110101ALI20200205BHEP

Ipc: A61B 8/08 20060101ALI20200205BHEP

Ipc: G06T 19/00 20110101ALI20200205BHEP

INTG Intention to grant announced

Effective date: 20200302

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE PATENT HAS BEEN GRANTED

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: AT

Ref legal event code: REF

Ref document number: 1307866

Country of ref document: AT

Kind code of ref document: T

Effective date: 20200915

Ref country code: CH

Ref legal event code: EP

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 602014069774

Country of ref document: DE

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: LT

Ref legal event code: MG4D

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20201202

Ref country code: HR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: LT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: NO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20201202

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20201203

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

REG Reference to a national code

Ref country code: NL

Ref legal event code: MP

Effective date: 20200902

REG Reference to a national code

Ref country code: AT

Ref legal event code: MK05

Ref document number: 1307866

Country of ref document: AT

Kind code of ref document: T

Effective date: 20200902

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: RS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: LV

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210104

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: SM

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: AL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: IS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210102

REG Reference to a national code

Ref country code: DE

Ref legal event code: R097

Ref document number: 602014069774

Country of ref document: DE

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

26N No opposition filed

Effective date: 20210603

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

Ref country code: SI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

REG Reference to a national code

Ref country code: DE

Ref legal event code: R119

Ref document number: 602014069774

Country of ref document: DE

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

GBPC Gb: european patent ceased through non-payment of renewal fee

Effective date: 20210704

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

REG Reference to a national code

Ref country code: BE

Ref legal event code: MM

Effective date: 20210731

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20210731

Ref country code: GB

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20210704

Ref country code: DE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220201

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20210731

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20210704

Ref country code: FR

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20210731

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20210704

Ref country code: BE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20210731

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: HU

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO

Effective date: 20140704

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20200902