US9576366B2 - Tracking system and tracking method using the same - Google Patents
Tracking system and tracking method using the same Download PDFInfo
- Publication number
- US9576366B2 US9576366B2 US14/370,549 US201414370549A US9576366B2 US 9576366 B2 US9576366 B2 US 9576366B2 US 201414370549 A US201414370549 A US 201414370549A US 9576366 B2 US9576366 B2 US 9576366B2
- Authority
- US
- United States
- Prior art keywords
- markers
- image
- image forming
- forming unit
- dimensional coordinates
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
- G06T7/74—Determining position or orientation of objects or cameras using feature-based methods involving reference images or patches
-
- G06T7/0044—
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/313—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor for introducing through surgical openings, e.g. laparoscopes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/30—Surgical robots
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
-
- H04N13/0271—
-
- H04N13/0296—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/271—Image signal generators wherein the generated image signals comprise depth maps or disparity maps
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/296—Synchronisation thereof; Control thereof
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2055—Optical tracking systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2055—Optical tracking systems
- A61B2034/2057—Details of tracking cameras
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B2090/363—Use of fiducial points
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3937—Visible markers
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3937—Visible markers
- A61B2090/3945—Active visible markers, e.g. light emitting diodes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3983—Reference marker arrangements for use with image guided surgery
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30204—Marker
Definitions
- Exemplary embodiments of the present invention relate to a tracking system and tracking method using the same. More particularly, exemplary embodiments of the present invention relate to a tracking system and tracking method using the same for surgery capable of detecting a spatial and a direction information of a target by tracking coordinates of markers attached on the target, in which the target are markers attached on a patient or a surgical instrument.
- a navigation system is used to navigate to an exact lesion of a patient by tracking and detecting a spatial position and a direction of a target such as lesion portion or surgical instrument.
- the navigation system described above includes a tracking system which is capable of tracking and detecting a spatial position and direction of a target such as lesion or surgical instrument.
- the tracking system described above includes a plurality of markers attached on a lesion or a surgical instrument, a first and second image forming units to form images of lights emitted from the markers, and a processor calculating three-dimensional coordinates of the markers which are connected to the first and second image forming units and calculating a spatial position and a direction of the target by comparing pre-stored information of straight lines connecting the markers adjacent to each other and angle information formed by a pair of straight lines adjacent to each other to the three-dimensional coordinates of the markers.
- a trigonometry is used in an assumption that a coordinate of marker which is emitted from one marker and formed image in a first image forming unit and a coordinate of marker which is emitted from one marker and formed image in a second image forming unit are identical.
- the technical problem of the present invention is to provide a tracking system and method using the same capable of reducing a manufacturing cost as well as minimizing a restriction of a surgical space by achieving compact of a system through calculating three-dimensional coordinates of each of markers by using only one image forming unit.
- a tracking system includes at least three markers which are attached on a target to emit lights or reflect lights emitted from a light source, a reflector which reflects light emitted from the makers or reflected light from the markers, an image forming unit which forms a direct image by directly receiving lights emitted from the markers and, at the same time, forming a reflection image by receiving lights reflected from the reflector after being emitted from markers, and a processor which calculates three-dimensional coordinates of each markers by using the direct image and the reflection image formed on the image forming unit, compares the three-dimensional coordinates of the markers to pre-stored geometric information of the markers which are adjacent to each other, and calculates a spatial position and a direction of the target.
- the reflector may be a mirror to form a reflection image by reflecting the emitted light from the markers to the image forming unit.
- the reflector is positioned on the same optical path as the image forming unit, changes at least one of a position, an angle and a shape of a reflection surface under the control of the processor, and changes an image forming position of the reflection image.
- the image forming unit may be a camera capable of forming image by receiving a direct light emitted from the markers and a reflected light reflected by the reflector.
- geometric information of the markers may be length information coupling markers adjacent to each other and angle information formed by a pair of straight lines adjacent to each other.
- a tracking method includes forming a direct image on an image forming units by directly receiving lights emitted from at least three markers which are attached on a target and, at the same time, forming a reflection image on the image forming unit by receiving lights which are reflected from the reflector, the reflector is installed on a specific position and reflects lights emitted from the markers, calculating three-dimensional coordinates of each markers by using the direct image and the reflection image of the markers formed on the image forming unit through the processor, and calculating a spatial position and a direction of the target by comparing the three-dimensional coordinates of the each markers to pre-stored geometric information of markers which are adjacent to each other.
- the geometric information of the markers may be length information coupling markers which are adjacent to each other and angle information formed by a pair of straight lines adjacent to each other.
- the process of calculating the three-dimensional coordinates of the marker may further include calculating two-dimensional coordinates of the direct image and reflection image which are formed on the image forming unit through the processor, and calculating the three-dimensional coordinates of the markers by using the two-dimensional coordinates of the direct image and reflection image of the markers.
- lights emitted from each of the markers are directly flowed to an image forming unit and, at the same time, lights reflected by the reflector are directly flowed to the image forming units.
- lights emitted from each of the markers are flowed to the image forming unit through two paths (first path: marker ⁇ image forming unit, second path: marker ⁇ reflector ⁇ image forming unit), an image sensor of the image forming unit forms images of the markers according to the two paths, and therefore, it is possible to calculate a spatial position and a direction of the markers attached on the target by using only one image forming unit.
- FIG. 1 is a schematic diagram of a tracking system according to an embodiment of the present invention
- FIG. 2 is an example diagram of markers attached on a target
- FIG. 3 is an example diagram explaining a position change of a reflection image when a position of the marker is changed on a same optical path as a lens;
- FIG. 4 is a block diagram explaining a tracking method according to an embodiment of the present invention.
- FIG. 5 is a block diagram explaining a method of calculating three-dimensional coordinates
- FIG. 6 is an example diagram of an image sensor of the image forming unit in which a coordinate of a direct image and coordinate of a reflection image is virtually divided;
- FIG. 7 is a diagram explaining a relationship between two-dimensional coordinates and three-dimensional coordinates of a real marker.
- first, second, third etc. may be used herein to describe various elements, components, regions, layers and/or sections, these elements, components, and/or sections should not be limited by these terms. These terms are only used to distinguish one element, component, region, layer or section from another region, layer or section. Thus, a first element, component, or section discussed below could be termed a second element, component, or section without departing from the teachings of the present invention.
- At least three markers are attached and three-dimensional coordinates of the markers are calculated, geometric information of markers adjacent to each, which are pre-stored in a processor, are compared to the three-dimensional coordinates of the markers through the processor, and therefore, it is capable of calculating a spatial position and a direction of a target such as a lesion or surgical instrument.
- a target such as a lesion or surgical instrument.
- FIG. 1 is a schematic diagram of a tracking system according to an embodiment of the present invention
- FIG. 2 is an example diagram of markers attached on a target
- FIG. 3 is an example diagram explaining a position change of a reflection image when a position of the marker is changed on a same optical path as a lens.
- a marker is shown in FIGS. 1 and 3 , however, in a tracking system according to an embodiment of the present invention, at least three markers 110 111 and 112 are attached on a target 200 such as a lesion or a surgical instrument.
- a tracking system includes at least three markers 110 111 and 112 , a reflector 120 , an image forming unit 130 , and a processor 140 .
- At least three markers 110 111 and 112 are attached on the target 200 such as a lesion or a surgical instrument.
- the at least three markers 110 111 and 112 are separated to each other in a predetermined interval, the markers are attached on the target 200 such as a lesion or a surgical instrument to form specific angles A 1 A 2 and A 3 which are formed by a pair of straight lines of the adjacent markers, in which the straight lines L 1 L 2 and L 3 virtually connect the adjacent markers 110 111 and 112 .
- geometric information between the markers 110 111 and 112 which are adjacent to each other in other words, length information of straight lines L 1 L 2 and L 3 which connect the markers 112 which are adjacent to each other and angle information A 1 A 2 and A 3 formed by a pair of the straight lines, are stored in a memory 141 of the processor 140 .
- the markers 110 111 and 112 may be attached on the target 200 such as a lesion and a surgical instrument in a triangle shape, and straight line information L 1 L 2 and L 3 forming sides of the triangle in which the markers are used as vertices and angle information A 1 A 2 and A 3 in which angles are formed by a pair of straight lines coupling each of the markers 110 111 and 112 may be pre-stored in the memory 141 of the processor 140 .
- the markers 110 111 and 112 may be an active markers which emit lights by itself. As described above, when an active marker is used for the markers 110 111 and 112 , a light source is not needed.
- the markers 110 111 and 112 may be passive markers which reflect a light emitted from at least one light source 150 .
- at least one light source 150 may be arranged close to the image forming unit 130 .
- a pair of light source 150 may be arranged on both sides of the image forming unit 130 .
- the reflector 120 reflects lights emitted from the markers 110 111 and 112 or lights reflected by the markers 110 111 and 112 .
- the reflector 120 reflects lights emitted from the markers 110 111 and 112 or re-reflects lights reflected by the maker 110 111 and 112 which are emitted from the light source 150 .
- the reflector may be a mirror which reflects lights emitted from the markers 110 111 and 112 to form an image of a reflection image on the image forming unit 130 .
- the reflector 120 may be a spherical mirror.
- a spherical mirror is used for the reflector 120 , since a ⁇ value is reduced when a marker changes its position on a same optical path AX 1 as the image forming unit 130 as shown in FIG. 3 , a position of a marker is calculated based on the ⁇ value.
- the reflector may change an image formation position of a reflection image formed on the image forming unit 130 .
- changing the position, the angle, or the shape of the reflection surface of the reflector may be controlled by the processor 140 which is interlocked in a wired/wireless network.
- the processor 140 is capable of changing an optical path as the reflection image by moving the reflector 120 when the optical path is screened by a stand of a surgical robot, an arm, a doctor or a nurse, etc.
- the processor decides that there is an obstacle in an optical path, and in such a case, the processor may send a control information to the reflector 120 such that the control information may move the reflector 120 to pre-stored value by controlling at least one of the installation position, the angle, the shape of the reflection surface of the reflector 120 .
- the spatial position and the direction of the reflector 120 and the changed spatial position and the direction of the reflector 120 may be stored in the memory 141 integrated in the processor 140 .
- the image forming unit 130 directly receives lights emitted from the markers 110 111 and 112 and forms an image and, at the same time, receives lights reflected from the reflector 120 which are emitted from the markers 110 111 112 and forms an image.
- the image forming unit 130 may be a camera which receives lights from the reflector 120 that are directly emitted from the markers 110 111 and 112 , and forms images.
- the image forming unit 130 may include a lens 131 in which lights emitted from the markers 110 111 and 112 and lights reflected from the reflector 120 pass through a focal, a body portion 132 arranged on a back portion of the lens 131 in which an image sensor 133 is integrated, the image sensor forms images by receiving lights emitted from the markers 110 111 and 112 and lights reflected from the reflector 120 .
- the processor 140 calculates three-dimensional coordinates of the markers 110 111 and 112 using the direct image and the reflection image formed on the image forming unit 130 , and is capable of calculating a spatial position and a direction of the target 200 such as a lesion or a surgical instrument by comparing the three-dimensional coordinates of the markers 110 111 and 112 with the pre-stored geometric information of the adjacent markers 110 111 and 112 .
- a memory 141 is integrated in the processor 140 .
- geometric information between the markers which are adjacent to each other in other words, length information of straight lines L 1 L 2 and L 3 which connect the marker adjacent to each other and angle information A 1 A 2 and A 3 which are formed by the pair of straight lines coupling the markers 110 111 and 112 adjacent to each other may be pre-stored the memory 141 integrated in the processor 140 .
- a spatial position and a direction of the reflector 120 may be pre-stored in the memory 141 integrated in the processor 140 .
- lights emitted from the markers 110 111 and 112 flow to the image forming unit 130 and a direct image is formed on the image forming unit 130 , as well as, lights reflected from the reflector 120 in which the lights emitted from the markers 110 111 and 112 flow to the reflector 120 and flow to the image forming unit 130 , and a reflection image is formed on the image forming units 130 , and therefore, even though one image forming unit 130 is used, there is the same effect as using additional image forming unit as shown in a dotted line on the left side of the reflector 120 in FIGS. 1 and 3 .
- FIGS. 1-7 a tracking process of a spatial position and a direction of a target using a tracking system according to an embodiment of the present invention is described below.
- FIG. 4 is a block diagram explaining a tracking method according to an embodiment of the present invention
- FIG. 5 is a block diagram explaining a method of calculating three-dimensional coordinates
- FIG. 6 is an example diagram of an image sensor of the image forming unit in which a coordinate of a direct image and coordinate of a reflection image is virtually divided
- FIG. 7 is a diagram explaining a relationship between two-dimensional coordinates and three-dimensional coordinates of a real marker.
- At least three markers 110 111 and 112 which are attached on the target 200 are activated making the markers 110 111 and 112 to emit light, or, at least one light source is activated to irradiated light toward the markers 110 111 and 112 which are attached on the target 200 such that the light is reflected and emitted by the markers 110 111 and 112 (S 11 ).
- the markers 110 111 and 112 are activated to emit lights.
- at least three passive (non-self-luminous) markers 110 111 and 112 are attached on the target 200 , at least one light source is activated to irradiated light toward the passive markers 110 111 and 112 which are attached on the target 200 such that the light is reflected and emitted by the passive markers 110 111 and 112 .
- Lights emitted from the at least three markers 110 111 and 112 are directly transferred to the image forming unit 130 and a direct image of each of the markers 110 111 and 112 are formed in the image forming unit 130 , and at the same time, lights reflected from the reflector 120 , in which the light emitted from the at least markers 110 111 and 112 are transferred to the reflector 120 , are transferred to the image forming unit 130 and a reflection image of each of the markers 110 111 and 112 are formed in the image forming unit 130 (S 120 ).
- a direct image is formed on the image forming unit 130 by directly receiving lights emitted from at least three markers 110 111 and 112 which are attached on the target 200 and, at the same time, a reflection image is formed on the image forming unit 130 by receiving lights reflected from the reflector 120 in which the light emitted from the at least markers 110 111 and 112 are transferred to the reflector 120 .
- lights emitted from the at least three markers 110 111 and 112 are directly transferred to the image forming unit 130 through a first optical path, pass a lens of the image forming unit 130 , and make a direct image of the at least three markers 110 111 and 112 on an image sensor 133 integrated in the body portion 132 and, at the same time, lights emitted from the at least three markers 110 111 and 112 are reflected by the reflector 120 through a second optical path and transferred to the image forming unit 130 , pass a lens of the image forming unit 130 , and make a reflection image of the at least three markers 110 111 and 112 on an image sensor 133 integrated in the body portion 132 .
- lights emitted from the markers 110 111 and 112 are transferred to the image forming unit 130 in two paths (first path: marker ⁇ image forming unit, second path: marker ⁇ reflector ⁇ image forming unit), and make two images (direct image and reflection image) of each of the markers 110 111 and 112 of the two paths (first and second paths).
- FIG. 5 shows a detailed process of calculating three-dimensional coordinates of each of the markers 110 111 and 112 .
- a camera calibration is processed for each coordinates (S 132 ).
- three-dimensional coordinates of each of the markers 110 111 and 112 are calculated by using two-dimensional coordinates of the direct and reflection images (S 133 ).
- one side of the image sensor 133 is virtually divided in a FOV (field of view) of a direct image and another side of the image sensor is virtually divided in a FOV (field of view) of a reflection image
- two-dimensional coordinates of the direct image of the image sensor 133 is represented by a coordinate system (U,V)
- two-dimensional coordinates of the reflection image of the image sensor 133 is represented by a coordinate system (U′,V′).
- a relationship between the two-dimensional coordinates of the markers 110 111 and 112 in real space and the three-dimensional coordinates of the markers 110 111 and 112 in real space may be represented in a formula below.
- m is two dimensional coordinates of the markers in the image
- M is three-dimensional coordinates of the markers in real space
- A(R, t) is a matrix of the camera
- P 1 is a camera matrix of the direct image
- P 2 is a camera matrix of the reflection image
- P jT is a row vector of the matrix P.
- the formula 3 may be represented in Formula 4.
- W may be a scale factor
- Three-dimensional coordinates of the markers 110 111 and 112 are obtained by calculating X, Y, and Z through solving the linear equation represented in formula 4.
- three-dimensional coordinates of the markers 110 111 and 112 in real space are compared to pre-stored geometric information of the markers which are adjacent to each other through the processor 140 , and a spatial position and a direction of the markers 110 111 and 112 attached on the target 200 is calculated (S 140 ).
- geometric information between the adjacent markers 110 111 and 112 may be length information of straight lines L 1 L 2 and L 3 which connect the marker adjacent to each other and angle information A 1 A 2 and A 3 which are formed by the pair of straight lines coupling the markers 110 111 and 112 adjacent to each other.
- the spatial position and the direction of the markers 110 111 and 112 are calculated by comparing the three-dimensional coordinates of the markers 110 111 and 112 in real space to length information of straight lines L 1 L 2 and L 3 which connect the marker adjacent to each other and angle information A 1 A 2 and A 3 which are formed by the pair of straight lines coupling the markers 110 111 and 112 adjacent to each other, in which the length information and the angle information are pre-stored in the processor 140 .
- lights emitted from each of the markers 110 111 and 112 are directly flowed to an image forming unit 130 and a direct image is formed on the image forming unit 130 and, at the same time, lights reflected from the reflector 120 are flowed to the image forming unit 130 and a reflection image is formed on it.
- lights emitted from the markers 110 111 and 112 are transferred to the image forming unit 130 in two paths (first path: marker ⁇ image forming unit, second path: marker ⁇ reflector ⁇ image forming unit), and make two images (direct image and reflection image) of each of the markers 110 111 and 112 of the two paths (first and second paths).
- one image forming unit 130 is used to calculate a spatial position and direction of the markers 110 111 and 120 attached on the target 200 .
Abstract
Description
x L =P 1 X
x R =P 2 X [Formula 2]
x(P 3T X)−(P 1T X)=0
y(P 3T X)−(P 2T X)=0
x(P 2T X)−y(P 1T X)=0 [Formula 3]
Claims (9)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020130002841A KR101371384B1 (en) | 2013-01-10 | 2013-01-10 | Tracking system and method for tracking using the same |
KR10-2013-0002841 | 2013-01-10 | ||
PCT/KR2014/000131 WO2014109520A1 (en) | 2013-01-10 | 2014-01-07 | Tracking system and method for tracking by using same |
Publications (2)
Publication Number | Publication Date |
---|---|
US20150302590A1 US20150302590A1 (en) | 2015-10-22 |
US9576366B2 true US9576366B2 (en) | 2017-02-21 |
Family
ID=50647854
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/370,549 Active 2034-01-17 US9576366B2 (en) | 2013-01-10 | 2014-01-07 | Tracking system and tracking method using the same |
Country Status (6)
Country | Link |
---|---|
US (1) | US9576366B2 (en) |
EP (1) | EP2944285B1 (en) |
JP (1) | JP6147360B2 (en) |
KR (1) | KR101371384B1 (en) |
CN (1) | CN104902840B (en) |
WO (1) | WO2014109520A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160270860A1 (en) * | 2013-01-18 | 2016-09-22 | Koh Young Technology Inc. | Tracking system and tracking method using the same |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101638477B1 (en) * | 2014-09-19 | 2016-07-11 | 주식회사 고영테크놀러지 | Optical tracking system and registration method for coordinate system in optical tracking system |
US10363667B2 (en) * | 2015-11-30 | 2019-07-30 | Autodesk, Inc. | Optical measurement of object location in three dimensions |
CN109186947B (en) * | 2018-09-26 | 2019-11-12 | 福州大学 | A kind of electric appliance Three-Dimensional Dynamic characteristic Auto-Test System and method |
FR3103097B1 (en) * | 2019-11-19 | 2021-11-05 | Quantum Surgical | Navigation method for positioning a medical robot |
Citations (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5394875A (en) | 1993-10-21 | 1995-03-07 | Lewis; Judith T. | Automatic ultrasonic localization of targets implanted in a portion of the anatomy |
US5923417A (en) | 1997-09-26 | 1999-07-13 | Northern Digital Incorporated | System for determining the spatial position of a target |
WO2002002028A1 (en) | 2000-07-06 | 2002-01-10 | Synthes Ag Chur | Method and device for impingement detection |
US20020130977A1 (en) * | 2001-03-12 | 2002-09-19 | Taketoshi Hibi | Image projection apparatus |
US20030209096A1 (en) | 2001-01-30 | 2003-11-13 | Z-Kat, Inc. | Tool calibrator and tracker system |
US20040160594A1 (en) * | 2003-02-14 | 2004-08-19 | Metronor Asa | Measurement of spatial coordinates |
US20040223139A1 (en) * | 2003-05-09 | 2004-11-11 | Michael Vogel | Method for determining the spatial location and position of a reflector rod in relation to a marked ground point |
US20050015005A1 (en) | 2003-04-28 | 2005-01-20 | Kockro Ralf Alfons | Computer enhanced surgical navigation imaging system (camera probe) |
US20060082789A1 (en) * | 2004-10-05 | 2006-04-20 | Gunter Goldbach | Positional marker system with point light sources |
US20060192921A1 (en) * | 2005-02-25 | 2006-08-31 | Frieder Loesel | Device and method for aligning an eye with a surgical laser |
JP2006284442A (en) | 2005-04-01 | 2006-10-19 | Shimadzu Corp | Method of monitoring object and motion tracker using this |
US20060240568A1 (en) * | 2005-04-22 | 2006-10-26 | Petruno Patrick T | Assay test strips and reading same |
JP2007130398A (en) | 2005-11-14 | 2007-05-31 | Toshiba Corp | Optical position measuring unit |
US20070183041A1 (en) | 2006-02-09 | 2007-08-09 | Northern Digital Inc. | Retroreflective marker-tracking systems |
CN101049229A (en) | 2006-04-07 | 2007-10-10 | 株式会社拓普康 | Ophthalmic device |
KR100835186B1 (en) | 2007-01-29 | 2008-06-04 | 주식회사 프리진 | System for classifying marker of motion capture using active marker |
US20090009759A1 (en) * | 2006-05-12 | 2009-01-08 | Vadim Backman | Systems, methods and apparatuses of elastic light scattering spectroscopy and low coherence enhanced backscattering spectroscopy |
US20090131921A1 (en) * | 2007-09-06 | 2009-05-21 | Lensx Lasers, Inc. | Precise Targeting of Surgical Photodisruption |
US20090137952A1 (en) * | 2007-08-14 | 2009-05-28 | Ramamurthy Bhaskar S | Robotic instrument systems and methods utilizing optical fiber sensor |
JP2009294048A (en) | 2008-06-04 | 2009-12-17 | Yokohama Rubber Co Ltd:The | Behavior measuring device of moving body, and behavior measuring method of moving body |
US20100188401A1 (en) * | 2005-08-26 | 2010-07-29 | Sony Corporation | Motion capture using primary and secondary markers |
EP2233099A2 (en) | 2009-03-24 | 2010-09-29 | Masmec S.r.L. | Computer-assisted system for guiding a surgical instrument during percutaneous diagnostic or therapeutic operations |
US20110211066A1 (en) | 2010-02-26 | 2011-09-01 | Canon Kabushiki Kaisha | Position and orientation measurement apparatus, position and orientation measurement method, and storage medium |
KR20110118640A (en) | 2008-12-31 | 2011-10-31 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | Configuration marker design and detection for instrument tracking |
US20120330154A1 (en) | 2010-03-04 | 2012-12-27 | Paul Beasley | Medical examination and/or treatment device |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6973202B2 (en) * | 1998-10-23 | 2005-12-06 | Varian Medical Systems Technologies, Inc. | Single-camera tracking of an object |
DE102006010767B4 (en) * | 2006-03-08 | 2008-04-17 | Carl Zeiss Surgical Gmbh | microscopy system |
-
2013
- 2013-01-10 KR KR1020130002841A patent/KR101371384B1/en active IP Right Grant
-
2014
- 2014-01-07 JP JP2015552571A patent/JP6147360B2/en active Active
- 2014-01-07 US US14/370,549 patent/US9576366B2/en active Active
- 2014-01-07 WO PCT/KR2014/000131 patent/WO2014109520A1/en active Application Filing
- 2014-01-07 EP EP14737829.3A patent/EP2944285B1/en active Active
- 2014-01-07 CN CN201480004280.1A patent/CN104902840B/en active Active
Patent Citations (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5394875A (en) | 1993-10-21 | 1995-03-07 | Lewis; Judith T. | Automatic ultrasonic localization of targets implanted in a portion of the anatomy |
US5923417A (en) | 1997-09-26 | 1999-07-13 | Northern Digital Incorporated | System for determining the spatial position of a target |
JP2001506762A (en) | 1997-09-26 | 2001-05-22 | ノーザン・デジタル・インコーポレーテッド | System for calculating the spatial position of a target |
US20030176783A1 (en) | 2000-07-06 | 2003-09-18 | Qingmao Hu | Method and device for impingement detection |
WO2002002028A1 (en) | 2000-07-06 | 2002-01-10 | Synthes Ag Chur | Method and device for impingement detection |
US20030209096A1 (en) | 2001-01-30 | 2003-11-13 | Z-Kat, Inc. | Tool calibrator and tracker system |
US20020130977A1 (en) * | 2001-03-12 | 2002-09-19 | Taketoshi Hibi | Image projection apparatus |
US20040160594A1 (en) * | 2003-02-14 | 2004-08-19 | Metronor Asa | Measurement of spatial coordinates |
US20050015005A1 (en) | 2003-04-28 | 2005-01-20 | Kockro Ralf Alfons | Computer enhanced surgical navigation imaging system (camera probe) |
US20040223139A1 (en) * | 2003-05-09 | 2004-11-11 | Michael Vogel | Method for determining the spatial location and position of a reflector rod in relation to a marked ground point |
US20060082789A1 (en) * | 2004-10-05 | 2006-04-20 | Gunter Goldbach | Positional marker system with point light sources |
US20060192921A1 (en) * | 2005-02-25 | 2006-08-31 | Frieder Loesel | Device and method for aligning an eye with a surgical laser |
JP2006284442A (en) | 2005-04-01 | 2006-10-19 | Shimadzu Corp | Method of monitoring object and motion tracker using this |
US20060240568A1 (en) * | 2005-04-22 | 2006-10-26 | Petruno Patrick T | Assay test strips and reading same |
US20100188401A1 (en) * | 2005-08-26 | 2010-07-29 | Sony Corporation | Motion capture using primary and secondary markers |
JP2007130398A (en) | 2005-11-14 | 2007-05-31 | Toshiba Corp | Optical position measuring unit |
CN101379412A (en) | 2006-02-09 | 2009-03-04 | 北方数字股份有限公司 | Retroreflective marker-tracking systems |
US20070183041A1 (en) | 2006-02-09 | 2007-08-09 | Northern Digital Inc. | Retroreflective marker-tracking systems |
CN101049229A (en) | 2006-04-07 | 2007-10-10 | 株式会社拓普康 | Ophthalmic device |
US20070236661A1 (en) | 2006-04-07 | 2007-10-11 | Yasufumi Fukuma | Opthalmologic Apparatus |
US20090009759A1 (en) * | 2006-05-12 | 2009-01-08 | Vadim Backman | Systems, methods and apparatuses of elastic light scattering spectroscopy and low coherence enhanced backscattering spectroscopy |
KR100835186B1 (en) | 2007-01-29 | 2008-06-04 | 주식회사 프리진 | System for classifying marker of motion capture using active marker |
US20090137952A1 (en) * | 2007-08-14 | 2009-05-28 | Ramamurthy Bhaskar S | Robotic instrument systems and methods utilizing optical fiber sensor |
US20090131921A1 (en) * | 2007-09-06 | 2009-05-21 | Lensx Lasers, Inc. | Precise Targeting of Surgical Photodisruption |
JP2009294048A (en) | 2008-06-04 | 2009-12-17 | Yokohama Rubber Co Ltd:The | Behavior measuring device of moving body, and behavior measuring method of moving body |
KR20110118640A (en) | 2008-12-31 | 2011-10-31 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | Configuration marker design and detection for instrument tracking |
EP2233099A2 (en) | 2009-03-24 | 2010-09-29 | Masmec S.r.L. | Computer-assisted system for guiding a surgical instrument during percutaneous diagnostic or therapeutic operations |
US20110211066A1 (en) | 2010-02-26 | 2011-09-01 | Canon Kabushiki Kaisha | Position and orientation measurement apparatus, position and orientation measurement method, and storage medium |
JP2011179910A (en) | 2010-02-26 | 2011-09-15 | Canon Inc | Device and method for measuring position and attitude, and program |
US20120330154A1 (en) | 2010-03-04 | 2012-12-27 | Paul Beasley | Medical examination and/or treatment device |
Non-Patent Citations (1)
Title |
---|
Written Opinion of the International Searching Authority for International Application No. PCT/KR2014/000131, dated Apr. 30, 2014. |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160270860A1 (en) * | 2013-01-18 | 2016-09-22 | Koh Young Technology Inc. | Tracking system and tracking method using the same |
Also Published As
Publication number | Publication date |
---|---|
CN104902840B (en) | 2017-08-01 |
JP2016502905A (en) | 2016-02-01 |
JP6147360B2 (en) | 2017-06-14 |
US20150302590A1 (en) | 2015-10-22 |
KR101371384B1 (en) | 2014-03-07 |
EP2944285B1 (en) | 2017-10-18 |
EP2944285A1 (en) | 2015-11-18 |
WO2014109520A1 (en) | 2014-07-17 |
CN104902840A (en) | 2015-09-09 |
EP2944285A4 (en) | 2016-08-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9576366B2 (en) | Tracking system and tracking method using the same | |
US8885177B2 (en) | Medical wide field of view optical tracking system | |
US20220395159A1 (en) | Device and method for assisting laparoscopic surgery - directing and maneuvering articulating tool | |
US20160270860A1 (en) | Tracking system and tracking method using the same | |
EP3076892B1 (en) | A medical optical tracking system | |
US20210038323A1 (en) | Optical tracking system and tracking method using the same | |
US20220175464A1 (en) | Tracker-Based Surgical Navigation | |
US11883105B2 (en) | Surgical navigation system using image segmentation | |
US7800643B2 (en) | Image obtaining apparatus | |
US11045259B2 (en) | Surgical navigation system | |
KR20140139698A (en) | Optical tracking system | |
EP2959857A1 (en) | Tracking system and tracking method using same | |
US8244495B2 (en) | Method and system for region of interest calibration parameter adjustment of tracking systems | |
US20220079687A1 (en) | Robot mounted camera registration and tracking system for orthopedic and neurological surgery | |
AU2023200152A1 (en) | Robot mounted camera registration and tracking system for orthopedic and neurological surgery | |
US11389250B2 (en) | Position detection system by fiber Bragg grating based optical sensors in surgical fields | |
Lukin | The new schemes of formation bistatic laser guide star |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: KYUNGPOOK NATIONAL UNIVERSITY INDUSTRY-ACADEMIC CO Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HONG, JONG-KYU;LEE, HYUN-KI;KIM, MIN-YOUNG;AND OTHERS;REEL/FRAME:033274/0906 Effective date: 20140630 Owner name: KOH YOUNG TECHNOLOGY INC., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HONG, JONG-KYU;LEE, HYUN-KI;KIM, MIN-YOUNG;AND OTHERS;REEL/FRAME:033274/0906 Effective date: 20140630 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.) |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |