CA1206617A - Optical position and orientation measurement techniques - Google Patents
Optical position and orientation measurement techniquesInfo
- Publication number
- CA1206617A CA1206617A CA000459133A CA459133A CA1206617A CA 1206617 A CA1206617 A CA 1206617A CA 000459133 A CA000459133 A CA 000459133A CA 459133 A CA459133 A CA 459133A CA 1206617 A CA1206617 A CA 1206617A
- Authority
- CA
- Canada
- Prior art keywords
- light sources
- light
- position sensor
- processor means
- position sensitive
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S5/00—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations
- G01S5/16—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using electromagnetic waves other than radio waves
- G01S5/163—Determination of attitude
Abstract
ABSTRACT OF THE DISCLOSURE
A non-contact sensing system for monitoring the position and orientatin of a rigid body in space having at least three distinct point light sources mounted thereon. At least two position sensor heads are fixedly mounted within detection range of light from the light sources, each position sensor head having a position sensitive detector. The lights are illuminated one at a time and in sequential arrangement and a processor computes the position in three-dimensional space of an illuminated light source with data provided simultaneously by both position sensitive detectors. The processor means also computes the positin of the object in all six degrees of freedom using the three-dimensional positional data of all three light sources.
A non-contact sensing system for monitoring the position and orientatin of a rigid body in space having at least three distinct point light sources mounted thereon. At least two position sensor heads are fixedly mounted within detection range of light from the light sources, each position sensor head having a position sensitive detector. The lights are illuminated one at a time and in sequential arrangement and a processor computes the position in three-dimensional space of an illuminated light source with data provided simultaneously by both position sensitive detectors. The processor means also computes the positin of the object in all six degrees of freedom using the three-dimensional positional data of all three light sources.
Description
The invention relates ~o a non-contact sensing system for monitoring the position and orientation of a rigid body in space. More specifically, the invention relates to such a system which can monitor the position and orientation of a pilot's helmet within the cockpit of a high performance tactical aircraf-t simulator.
In order for aircraft simulators to provide a realistic simulation of aircraft flight, it is necessary to display a simulated surrounding environment to the simulator pilot. In currently available flight simulators, the displays are provided on the windscreens in the cockpit areas. These systems require one CRT for each such wind-screen. With such a system, images of equal quality must be generated in each CRT regardless of where the pilot is looking. This is costly both in terms of hardware and computer time.
However, by projecting the image directly onto the visor of a pilot's helmet, a significant cost saving can be realized since the field of view is greatly reduced so that both the quality of the image (i.e. resolution) and the frame rate can be reduced. The field of view would be head slaved in order to provide an image which is instantaneously limited to the pilot's viewing area but which can be redirected to permit the pilot to "look"
in any direction. It is also contemplated to track eye position or point of regard and provide a higher resolution image slaved to the eye's foveal area. In such a system, the pilot would perceive high resolution imagery wherever he looks. However, at any instant, the system is only required to produce high resolution images over a 15 cone.
In order for aircraft simulators to provide a realistic simulation of aircraft flight, it is necessary to display a simulated surrounding environment to the simulator pilot. In currently available flight simulators, the displays are provided on the windscreens in the cockpit areas. These systems require one CRT for each such wind-screen. With such a system, images of equal quality must be generated in each CRT regardless of where the pilot is looking. This is costly both in terms of hardware and computer time.
However, by projecting the image directly onto the visor of a pilot's helmet, a significant cost saving can be realized since the field of view is greatly reduced so that both the quality of the image (i.e. resolution) and the frame rate can be reduced. The field of view would be head slaved in order to provide an image which is instantaneously limited to the pilot's viewing area but which can be redirected to permit the pilot to "look"
in any direction. It is also contemplated to track eye position or point of regard and provide a higher resolution image slaved to the eye's foveal area. In such a system, the pilot would perceive high resolution imagery wherever he looks. However, at any instant, the system is only required to produce high resolution images over a 15 cone.
2~
One of the difficulties associated with the use of helmet mounted visor displays is the maintenance of image stability. In order for the computer generated image to appear stationary as the pilot moves his head, the positionof the pilot's helmet must be kno~n relative to the simulator's fixed axis. This position is defined by three translation parameters tX, Y, Z) and three rotation parameters (Euler Angles). This information is also necessary in order to align the computer generated image with the frame of the windscreen. Thus, a ~helmet) position sensing system is needed.
Curxently available non-contact position sensing systems comprise a magnetic system as described in Raab, F.H.,Blood, E.B., Steiner, T.O. t and Jones, H.R., "Magnetic Position and Orientation Tracking System", IEEE Transactions on Aerospace and Electronic Systems, Vol AES-15, Mo. 5, September 1979, pp. 709-718, and a V-slit system as described in Lewis, C.J.G., "Helmet Mounted Display and Sigh-t Development", Proceedings of the Arnerican Helicopter Society, May 1979, pp. 79.17.1-79.17.13. Ho~ever, both of these systerns fail to meet required specifications.
It is therefore an object of the invention to provide a non-contact position sensing system which meets the required specifications.
In accordance with the invention, the system monitors the position orientation of a body in space which has at least three distinct point light sources mounted thereon. The system includes at least two position sensor heads each of which includes a position sensitive aetector .
The invention will be better understood by an examination ofthe fol].owing description, together with the accompanying drawings, in which:
FIGURE 1 is a schematic diagram of the inventive system~
FIGURE 2A is a schematic diagram of a position sensitive detector;
FIGURE 2B is an electron flow diagram oE the position sensitive detector; and FIGURE 3 illustrates an analog position processing circuit for processing data of the position sensiti~e detector.
Referring to Figure 1, 1 is a rigid object in space whose position and orientation must be monitored.
In th~ particular embodiment illustrated, 1 is a helmet which is worn by a pilot in the cockpit of a high perform-ance tactical aircraft simulator.
A mounting plate 3 carries at least three point light sources, preferably LED's, and the mountin~ plate is mounted on the helmet. Obviously, the LED's could be mounted directly on the helmet instead of on a mounting plate.
Fixedly mounted in the cockpit are position sensor heads 5 and 7. The sensorheadsare mounted within detection range of the li~ht from the LEDs, and are directed at the LEDs. Camera lenses 6 and 8 are disposed in front of the position sensor heads for reasons explained be].ow.
As the position sensor head is sensitive to a wide spectrum oE light, bandpass interference filters, illus-trated schematically at 9 and 11, are placed in front of each lens to minimize the effect of ambient light. The filters are spectrally matched to the light of the LEDs.
Each position sensor head includes a positionsensitive detector which will be discussed in association with Figures 2 and 3 below.
The output of each position sensor head is fed to an analog-to-digital converter 13, and the digital data is then fed to a computer 15. One output o~ 15 is fed to a controller 17 which provides a host computer with the position and orientation information of the helmet as computed in the computer.
As is known, and will be discussed below, at least three LEDs coupled with two position sensor heads are needed to determine the position and orientation of the helmet in all six degrees of freedom. In accordance with the invention, the LEDs are illuminated one at a time, and in sequential arrangement, and the three dimensional position of each LED is determined by a triangulation scheme using the data from the two position sensor heads~
It is also necessary to maintain a constant light intensity of each LED regardless of the position or orientation of the helmet.
LED controller l9 maintains a constant light intensity of the LEDs and also illuminates the LEDs one at a time and in sequential arrangement. As can be seen, the controller l9 is controlled by the computer lS which determines the re~uired LED current to maintain a constant light intensity as a function of the position of the helmet. The computer also selects the particular LED to be illuminated, and takes into account the fact that it was this LED illuminated when computing the position and orientation of the helmet.
~2q~
Turning now to Figures 2A and 2B, as can be seen the positionsensitive detector is a planar photodiode 21 with very uniform resistive layers formed over both the top and bottom surfaces. The camera lenses 6 and 8 are used to focus the infrared light from the LEDs onto the active area of this detector.
This results in the generation of electron-hole pairs inthe depletion layer directly under the liyht spot.
The electrons migr~te to the N-layer where they are channelled between two electrodes. Since the N-layer has a uniform resistivity, the current flowing to the pair of Y electrodes will depend on its distance from the point of incidence of the light.
The same basic principles apply to the P-layer, i.e., holes will migrate to the P-layer. The current signals are then processed using, for example, the circuitry illustrated in Figure 3.
Turning now to Figure 3, the processing circuit comprises preamps 23 connected to each one of the X and Y
electrode pairs. Adders 25A add the contents of the X
electrodes and adders 25B add the contents of the Y
electrodes. Subtractors 27A subtract the contents of the X electrodes and subtractors 27B subtract the contents of the Y electrodes~ The outputs of 25A and 27A are fed to background cancellation circuits 29A and the outputs of 25B and 27B are fed to background cancellation circuits 29B, and the outputs of cancellation circuits 29A and 29B
are fed to dividers 3]A and 31B, respectively.
With filtering, as per filters 9 and 11, the resolution of the position sensor head can be as high as 0.02% o full scale. The filtering is necessary because of the high fre~uency jitter which is normally present.
With the dynamic performance of the inven-tive device, it can be used for vibration measurements at Erequencies of up to 50 kHz.
It is necessary to use two sensors viewing the same LED patterns (at least three LEDs) to uniquely deter-mine the helme-t position in all siY degrees of freedom.
The image on the sensitive area of the detector of each LED is a two-dimensional perspective projection of an LED
moving in three-dimensional space. Consequently, the three-dimensional position of the LED canno-t be recovered from its image coordinates. At best, only the direction of the vector between the LED and the sensor can be determined. If each LED is viewed from two locations, a triangulation scheme can be used to compute the position of each LED in three dimensions. Therefore, with a minimum of three LEDs and two sensors, the helmet position can be uniquely determined in all six degrees of freedom.
The addition of more LEDs results in an over-determined system of equations which can increase theaccuracy of the measurement through the application of least squares analysis as taught in Roach, J.W. and Aggarwal, J.K., "Determining the Movement of Objects from a Sequence of Images", IEEE Transactions on Pattern Analysis and Machine Intelligence, VolO PAMI-2, No. 6, November 1980, pp. 554-562.
Data processingtasksfor the optical helmet position sensing systems are accomplished, in one embodi-ment, by an INTEl iSBC 86/12 single board computer. This Board's capabilities are summari~ed as follows:
CPU SPEED 5MHz RAM (BASE BOARD) 32 K
EPROM (BASE BOARD) 16 K
SERIAL I/O
PARALLEL I/O I.INES 24 In order to determine the helmet position and orientation, a large number of floating point operations must be per~ormed~ Several trigometric functions are also required. The floating point operations can be performed by the 8087 Numeric ~rocessor. The 8087 has the ability to perform the high speed floating point mathematics required i~ complex control algorithms.
Helmet position and orientation data are, as above-mentioned, transferred from the computer to a host computer. In the preferred embodiment, the transfer from the INTEL 86/12 is to a SEL 32/55 host computer. In this embodiment, the control 17 comprises a MULTIBUS-SELBUS
Controller (MBSEL). The MBS~L board is an intelligent DMA controller which communicates with the SELBUS via a high speed data interface board (HSD). Because of the 86/12 board's dual port RAM, the MBSEL can access the position and orientation data directly via the MULTIBUS.
Data transfers are transparent to both the SEL and 8086 processors.
Again, in the particular embodiment, a custom-built analog board 13 is used to read the output signals from the position sensor heads and to convert them into digital signals. In order to minirnize the conversion time, two independent 14-bit analog-to-d:igital con~erters were included on this boardO
Methods for the mathematical analysis of the data are taught in, for example, Lenox, J.B., Six De~ree of Freedom Human Eyeball Movement Analysis Involving Steriometric Techniques, PhoD. Thesis, Stanford University, Stanford, California, 1976~ Schut, ~.H., "On Exact Linear Equations for the Computation of the Rotational Elements of Absolute Orientation", Photogrammetria, Vol. 17, No. 1, 1960, pp. 34-37, Schut, G.H., "Formation of Strips from Independent Models", AP-PR 36, NRC~9695, Division of Physics, National Research Councilof Canada, Ottawa, July 1967, and Thompson, E.H., "An Exact Linear Solution of the Problem of Absolute Orientation", Photogrammetria, Vol. 15, No. 4, 1959, pp. 163-179.
In comparison with other non-contact position sensing systems, the inventive system offers high resolution and speed at a relativel~ low cost. These features suggest other possible applications of the inventive system in the areas of robot vision, industrial control, kinesiology and laboratory instrumentation. Accordingly, the inventive system has uses outside of the environment as described above in the preferred embodiment.
Although a particular embodiment has been des-cribed, this was for the purpose of illustrating, but not limiting, the invention. Various modifications, which will come readily to the mind of one skilled in the art, are within the scope of the invention as defined in the appended claims.
One of the difficulties associated with the use of helmet mounted visor displays is the maintenance of image stability. In order for the computer generated image to appear stationary as the pilot moves his head, the positionof the pilot's helmet must be kno~n relative to the simulator's fixed axis. This position is defined by three translation parameters tX, Y, Z) and three rotation parameters (Euler Angles). This information is also necessary in order to align the computer generated image with the frame of the windscreen. Thus, a ~helmet) position sensing system is needed.
Curxently available non-contact position sensing systems comprise a magnetic system as described in Raab, F.H.,Blood, E.B., Steiner, T.O. t and Jones, H.R., "Magnetic Position and Orientation Tracking System", IEEE Transactions on Aerospace and Electronic Systems, Vol AES-15, Mo. 5, September 1979, pp. 709-718, and a V-slit system as described in Lewis, C.J.G., "Helmet Mounted Display and Sigh-t Development", Proceedings of the Arnerican Helicopter Society, May 1979, pp. 79.17.1-79.17.13. Ho~ever, both of these systerns fail to meet required specifications.
It is therefore an object of the invention to provide a non-contact position sensing system which meets the required specifications.
In accordance with the invention, the system monitors the position orientation of a body in space which has at least three distinct point light sources mounted thereon. The system includes at least two position sensor heads each of which includes a position sensitive aetector .
The invention will be better understood by an examination ofthe fol].owing description, together with the accompanying drawings, in which:
FIGURE 1 is a schematic diagram of the inventive system~
FIGURE 2A is a schematic diagram of a position sensitive detector;
FIGURE 2B is an electron flow diagram oE the position sensitive detector; and FIGURE 3 illustrates an analog position processing circuit for processing data of the position sensiti~e detector.
Referring to Figure 1, 1 is a rigid object in space whose position and orientation must be monitored.
In th~ particular embodiment illustrated, 1 is a helmet which is worn by a pilot in the cockpit of a high perform-ance tactical aircraft simulator.
A mounting plate 3 carries at least three point light sources, preferably LED's, and the mountin~ plate is mounted on the helmet. Obviously, the LED's could be mounted directly on the helmet instead of on a mounting plate.
Fixedly mounted in the cockpit are position sensor heads 5 and 7. The sensorheadsare mounted within detection range of the li~ht from the LEDs, and are directed at the LEDs. Camera lenses 6 and 8 are disposed in front of the position sensor heads for reasons explained be].ow.
As the position sensor head is sensitive to a wide spectrum oE light, bandpass interference filters, illus-trated schematically at 9 and 11, are placed in front of each lens to minimize the effect of ambient light. The filters are spectrally matched to the light of the LEDs.
Each position sensor head includes a positionsensitive detector which will be discussed in association with Figures 2 and 3 below.
The output of each position sensor head is fed to an analog-to-digital converter 13, and the digital data is then fed to a computer 15. One output o~ 15 is fed to a controller 17 which provides a host computer with the position and orientation information of the helmet as computed in the computer.
As is known, and will be discussed below, at least three LEDs coupled with two position sensor heads are needed to determine the position and orientation of the helmet in all six degrees of freedom. In accordance with the invention, the LEDs are illuminated one at a time, and in sequential arrangement, and the three dimensional position of each LED is determined by a triangulation scheme using the data from the two position sensor heads~
It is also necessary to maintain a constant light intensity of each LED regardless of the position or orientation of the helmet.
LED controller l9 maintains a constant light intensity of the LEDs and also illuminates the LEDs one at a time and in sequential arrangement. As can be seen, the controller l9 is controlled by the computer lS which determines the re~uired LED current to maintain a constant light intensity as a function of the position of the helmet. The computer also selects the particular LED to be illuminated, and takes into account the fact that it was this LED illuminated when computing the position and orientation of the helmet.
~2q~
Turning now to Figures 2A and 2B, as can be seen the positionsensitive detector is a planar photodiode 21 with very uniform resistive layers formed over both the top and bottom surfaces. The camera lenses 6 and 8 are used to focus the infrared light from the LEDs onto the active area of this detector.
This results in the generation of electron-hole pairs inthe depletion layer directly under the liyht spot.
The electrons migr~te to the N-layer where they are channelled between two electrodes. Since the N-layer has a uniform resistivity, the current flowing to the pair of Y electrodes will depend on its distance from the point of incidence of the light.
The same basic principles apply to the P-layer, i.e., holes will migrate to the P-layer. The current signals are then processed using, for example, the circuitry illustrated in Figure 3.
Turning now to Figure 3, the processing circuit comprises preamps 23 connected to each one of the X and Y
electrode pairs. Adders 25A add the contents of the X
electrodes and adders 25B add the contents of the Y
electrodes. Subtractors 27A subtract the contents of the X electrodes and subtractors 27B subtract the contents of the Y electrodes~ The outputs of 25A and 27A are fed to background cancellation circuits 29A and the outputs of 25B and 27B are fed to background cancellation circuits 29B, and the outputs of cancellation circuits 29A and 29B
are fed to dividers 3]A and 31B, respectively.
With filtering, as per filters 9 and 11, the resolution of the position sensor head can be as high as 0.02% o full scale. The filtering is necessary because of the high fre~uency jitter which is normally present.
With the dynamic performance of the inven-tive device, it can be used for vibration measurements at Erequencies of up to 50 kHz.
It is necessary to use two sensors viewing the same LED patterns (at least three LEDs) to uniquely deter-mine the helme-t position in all siY degrees of freedom.
The image on the sensitive area of the detector of each LED is a two-dimensional perspective projection of an LED
moving in three-dimensional space. Consequently, the three-dimensional position of the LED canno-t be recovered from its image coordinates. At best, only the direction of the vector between the LED and the sensor can be determined. If each LED is viewed from two locations, a triangulation scheme can be used to compute the position of each LED in three dimensions. Therefore, with a minimum of three LEDs and two sensors, the helmet position can be uniquely determined in all six degrees of freedom.
The addition of more LEDs results in an over-determined system of equations which can increase theaccuracy of the measurement through the application of least squares analysis as taught in Roach, J.W. and Aggarwal, J.K., "Determining the Movement of Objects from a Sequence of Images", IEEE Transactions on Pattern Analysis and Machine Intelligence, VolO PAMI-2, No. 6, November 1980, pp. 554-562.
Data processingtasksfor the optical helmet position sensing systems are accomplished, in one embodi-ment, by an INTEl iSBC 86/12 single board computer. This Board's capabilities are summari~ed as follows:
CPU SPEED 5MHz RAM (BASE BOARD) 32 K
EPROM (BASE BOARD) 16 K
SERIAL I/O
PARALLEL I/O I.INES 24 In order to determine the helmet position and orientation, a large number of floating point operations must be per~ormed~ Several trigometric functions are also required. The floating point operations can be performed by the 8087 Numeric ~rocessor. The 8087 has the ability to perform the high speed floating point mathematics required i~ complex control algorithms.
Helmet position and orientation data are, as above-mentioned, transferred from the computer to a host computer. In the preferred embodiment, the transfer from the INTEL 86/12 is to a SEL 32/55 host computer. In this embodiment, the control 17 comprises a MULTIBUS-SELBUS
Controller (MBSEL). The MBS~L board is an intelligent DMA controller which communicates with the SELBUS via a high speed data interface board (HSD). Because of the 86/12 board's dual port RAM, the MBSEL can access the position and orientation data directly via the MULTIBUS.
Data transfers are transparent to both the SEL and 8086 processors.
Again, in the particular embodiment, a custom-built analog board 13 is used to read the output signals from the position sensor heads and to convert them into digital signals. In order to minirnize the conversion time, two independent 14-bit analog-to-d:igital con~erters were included on this boardO
Methods for the mathematical analysis of the data are taught in, for example, Lenox, J.B., Six De~ree of Freedom Human Eyeball Movement Analysis Involving Steriometric Techniques, PhoD. Thesis, Stanford University, Stanford, California, 1976~ Schut, ~.H., "On Exact Linear Equations for the Computation of the Rotational Elements of Absolute Orientation", Photogrammetria, Vol. 17, No. 1, 1960, pp. 34-37, Schut, G.H., "Formation of Strips from Independent Models", AP-PR 36, NRC~9695, Division of Physics, National Research Councilof Canada, Ottawa, July 1967, and Thompson, E.H., "An Exact Linear Solution of the Problem of Absolute Orientation", Photogrammetria, Vol. 15, No. 4, 1959, pp. 163-179.
In comparison with other non-contact position sensing systems, the inventive system offers high resolution and speed at a relativel~ low cost. These features suggest other possible applications of the inventive system in the areas of robot vision, industrial control, kinesiology and laboratory instrumentation. Accordingly, the inventive system has uses outside of the environment as described above in the preferred embodiment.
Although a particular embodiment has been des-cribed, this was for the purpose of illustrating, but not limiting, the invention. Various modifications, which will come readily to the mind of one skilled in the art, are within the scope of the invention as defined in the appended claims.
Claims (8)
1. A non-contact sensing system for monitoring the position and orientation of a rigid body in space wherein at least three distinct point light sources are mounted on said body for movement therewith;
said system comprising:
at least two position sensor heads fixedly mounted within detection range of light from said light sources, said position sensor heads being directed at said light sources;
characterized in that each position sensor head includes a position sensitive detector;
means for illuminating said light sources one at a time and in a sequential arrangement;
processor means;
said position sensitive detector outputs being connected to the input of said processor means;
said processor means being adapted to compute the position in three-dimensional space of an illuminated light source with data concerning said illuminated light source provided simultaneously by both position sensitive detectors;
said processor means also being adapted to compute the position of said object in all six degrees of freedom using the three-dimensional positional data of all three light sources.
said system comprising:
at least two position sensor heads fixedly mounted within detection range of light from said light sources, said position sensor heads being directed at said light sources;
characterized in that each position sensor head includes a position sensitive detector;
means for illuminating said light sources one at a time and in a sequential arrangement;
processor means;
said position sensitive detector outputs being connected to the input of said processor means;
said processor means being adapted to compute the position in three-dimensional space of an illuminated light source with data concerning said illuminated light source provided simultaneously by both position sensitive detectors;
said processor means also being adapted to compute the position of said object in all six degrees of freedom using the three-dimensional positional data of all three light sources.
2. A system as defined in claim 1 wherein said position sensitive detectors provide analog signals at their outputs;
analog-to-digital conversion means between the outputs of said position sensitive detectors and the input of said processing means whereby to convert said analog signals into digital signals for use in said processor means.
analog-to-digital conversion means between the outputs of said position sensitive detectors and the input of said processing means whereby to convert said analog signals into digital signals for use in said processor means.
3. A system as defined in claim 2 and including means for varying power to said light sources whereby to maintain a constant light intensity from each light source at said position sensitive detectors regardless of the position and orientation of said rigid body.
4. A system as defined in claim 3 and further including lens means on each position sensor head whereby to focus light from said light sources onto a respective position sensitive detector.
5. A system as defined in claim 4 and including filter means on each lens, the filter means being spectrally matched to the light sources.
6. A system as defined in claim 5 wherein said rigid body comprises a pilot's helmet within the cockpit of an aircraft simulator;
said position sensor heads being mounted in the interior of said simulator.
said position sensor heads being mounted in the interior of said simulator.
7. A system as defined in claim 6 wherein said light sources comprise LEDs, said means for varying the light intensity comprising means for varying the current to said LEDs.
8. A non-contact method for monitoring the position and orientation of a rigid body in space having at least three distinct point light sources mounted thereon for movement therewith;
at least two position sensor heads fixedly mounted within detection range of light from said light sources, each position sensor head including a position sensitive detector, said position sensor heads being directed at said light sources;
and a processor means;
said method comprising;
illuminating said light sources one at a time and in sequential arrangement;
computing, by said processor means, the position in three-dimensional space of an illuminated light source with data concerning said illuminated light source pro-vided simultaneously by both position sensitive detectors to said processor means;
computing the position of said object in all six degrees of freedom using the three-dimensional positional data of all three light sources.
at least two position sensor heads fixedly mounted within detection range of light from said light sources, each position sensor head including a position sensitive detector, said position sensor heads being directed at said light sources;
and a processor means;
said method comprising;
illuminating said light sources one at a time and in sequential arrangement;
computing, by said processor means, the position in three-dimensional space of an illuminated light source with data concerning said illuminated light source pro-vided simultaneously by both position sensitive detectors to said processor means;
computing the position of said object in all six degrees of freedom using the three-dimensional positional data of all three light sources.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US613,063 | 1984-05-22 | ||
US06/613,063 US4649504A (en) | 1984-05-22 | 1984-05-22 | Optical position and orientation measurement techniques |
Publications (1)
Publication Number | Publication Date |
---|---|
CA1206617A true CA1206617A (en) | 1986-06-24 |
Family
ID=24455719
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA000459133A Expired CA1206617A (en) | 1984-05-22 | 1984-07-18 | Optical position and orientation measurement techniques |
Country Status (5)
Country | Link |
---|---|
US (1) | US4649504A (en) |
EP (1) | EP0162713B1 (en) |
JP (1) | JPS6134406A (en) |
CA (1) | CA1206617A (en) |
DE (1) | DE3579363D1 (en) |
Families Citing this family (206)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
IL77610A (en) * | 1986-01-15 | 1994-01-25 | Technion Res & Dev Foundation | Single camera three-dimensional head position sensing system |
SE457017B (en) * | 1986-02-04 | 1988-11-21 | Geotronics Ab | DEVICE FOR MEASURING THE DIMENSION OF SIGNIFICANT VERTICAL, CIRCUMSTANCES |
DE3606399A1 (en) * | 1986-02-27 | 1987-09-03 | Messerschmitt Boelkow Blohm | MEASURING DEVICE FOR DETERMINING THE POSITION OF AN OBJECT |
US4825393A (en) * | 1986-04-23 | 1989-04-25 | Hitachi, Ltd. | Position measuring method |
FR2723207B1 (en) * | 1986-06-17 | 1996-12-13 | Thomson Csf | SYSTEM DETERMINING THE ORIENTATION AND LOCATION OF A MOBILE BODY RELATIVE TO A STRUCTURE, ESPECIALLY USEFUL FOR A HELMET VIEWFINDER |
FR2723208B1 (en) * | 1986-06-17 | 1996-11-15 | Thomson Csf | SYSTEM FOR LOCATING A DIRECTION LINKED TO A MOBILE BODY RELATIVE TO A STRUCTURE, ESPECIALLY USEFUL FOR A HELMET VIEWFINDER |
FR2603974B1 (en) * | 1986-09-12 | 1988-11-04 | Thomson Csf | SUPPORT DEVICE SERVED BY THE MOVEMENT OF A MOBILE BODY RELATIVE TO A STRUCTURE, USEFUL FOR LARGE-FIELD HELMET VIEWFINDERS |
WO1988004060A2 (en) * | 1986-11-27 | 1988-06-02 | Starpeak Computers Limited | Locating system |
GB8715184D0 (en) * | 1987-06-29 | 1987-10-21 | Gec Avionics | Stereoscopic presentation of data |
US4841460A (en) * | 1987-09-08 | 1989-06-20 | Perceptron, Inc. | Method and apparatus for calibrating a non-contact gauging sensor with respect to an external coordinate system |
US5086404A (en) * | 1988-09-02 | 1992-02-04 | Claussen Claus Frenz | Device for simultaneous continuous and separate recording and measurement of head and body movements during standing, walking and stepping |
US4982188A (en) * | 1988-09-20 | 1991-01-01 | Grumman Aerospace Corporation | System for measuring positional characteristics of an ejected object |
US4932777A (en) * | 1988-09-30 | 1990-06-12 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Electro-optical spin measurement system |
GB2234877A (en) * | 1989-08-09 | 1991-02-13 | Marconi Gec Ltd | Determining orientation of pilot's helmet for weapon aiming |
FR2652928B1 (en) | 1989-10-05 | 1994-07-29 | Diadix Sa | INTERACTIVE LOCAL INTERVENTION SYSTEM WITHIN A AREA OF A NON-HOMOGENEOUS STRUCTURE. |
WO1991018258A1 (en) * | 1990-05-19 | 1991-11-28 | Kabushiki Kaisha Topcon | Method of tridimensional measuring, reference scale and self-illuminating reference scale for tridimensional measuring |
US5208641A (en) * | 1990-09-28 | 1993-05-04 | Honeywell Inc. | Laser cavity helmet mounted sight |
GB2251751A (en) * | 1990-10-09 | 1992-07-15 | Gaertner W W Res | Position and orientation measurement |
US5198877A (en) * | 1990-10-15 | 1993-03-30 | Pixsys, Inc. | Method and apparatus for three-dimensional non-contact shape sensing |
DE69132412T2 (en) | 1990-10-19 | 2001-03-01 | Univ St Louis | LOCALIZATION SYSTEM FOR A SURGICAL PROBE FOR USE ON THE HEAD |
US6347240B1 (en) | 1990-10-19 | 2002-02-12 | St. Louis University | System and method for use in displaying images of a body part |
US5059789A (en) * | 1990-10-22 | 1991-10-22 | International Business Machines Corp. | Optical position and orientation sensor |
US5187540A (en) * | 1990-10-31 | 1993-02-16 | Gec Ferranti Defence Systems Limited | Optical system for the remote determination of position and orientation |
WO1992009904A1 (en) * | 1990-11-29 | 1992-06-11 | Vpl Research, Inc. | Absolute position tracker |
US5128794A (en) * | 1990-12-31 | 1992-07-07 | Honeywell Inc. | Scanning laser helmet mounted sight |
US5227985A (en) * | 1991-08-19 | 1993-07-13 | University Of Maryland | Computer vision system for position monitoring in three dimensions using non-coplanar light sources attached to a monitored object |
FR2682798B1 (en) * | 1991-10-22 | 1994-01-21 | Sextant Avionique | METHOD AND DEVICE FOR OPTIMIZING THE PERFORMANCE OF A LIQUID CRYSTAL MATRIX SCREEN AS A FUNCTION OF THE ANGLE OF OBSERVATION. |
US5603318A (en) | 1992-04-21 | 1997-02-18 | University Of Utah Research Foundation | Apparatus and method for photogrammetric surgical localization |
US5339259A (en) * | 1992-07-10 | 1994-08-16 | Northrop Grumman Corporation | High speed high resolution ultrasonic position and orientation tracker |
US6757557B1 (en) | 1992-08-14 | 2004-06-29 | British Telecommunications | Position location system |
WO1994004938A1 (en) * | 1992-08-14 | 1994-03-03 | British Telecommunications Public Limited Company | Position location system |
DE4308753C1 (en) * | 1993-03-19 | 1994-07-21 | Deutsche Aerospace | Method and device for image-based position detection |
CA2161430C (en) | 1993-04-26 | 2001-07-03 | Richard D. Bucholz | System and method for indicating the position of a surgical probe |
US5297061A (en) * | 1993-05-19 | 1994-03-22 | University Of Maryland | Three dimensional pointing device monitored by computer vision |
US5396331A (en) * | 1993-08-10 | 1995-03-07 | Sanyo Machine Works, Ltd. | Method for executing three-dimensional measurement utilizing correctively computing the absolute positions of CCD cameras when image data vary |
US5572646A (en) * | 1993-08-25 | 1996-11-05 | Casio Computer Co., Ltd. | Apparatus for displaying images of living things to show growing and/or moving of the living things |
US5606494A (en) * | 1993-11-25 | 1997-02-25 | Casio Computer Co., Ltd. | Switching apparatus |
US5424556A (en) * | 1993-11-30 | 1995-06-13 | Honeywell Inc. | Gradient reflector location sensing system |
GB2284957B (en) * | 1993-12-14 | 1998-02-18 | Gec Marconi Avionics Holdings | Optical systems for the remote tracking of the position and/or orientation of an object |
US5436542A (en) * | 1994-01-28 | 1995-07-25 | Surgix, Inc. | Telescopic camera mount with remotely controlled positioning |
DE4412073A1 (en) * | 1994-04-08 | 1995-10-12 | Moeller J D Optik | Surgical microscope unit |
DE69531994T2 (en) | 1994-09-15 | 2004-07-22 | OEC Medical Systems, Inc., Boston | SYSTEM FOR POSITION DETECTION BY MEANS OF A REFERENCE UNIT ATTACHED TO A PATIENT'S HEAD FOR USE IN THE MEDICAL AREA |
GB2308186B (en) * | 1994-09-28 | 1999-06-09 | William Richard Fright | Arbitrary-geometry laser surface scanner |
US6978166B2 (en) * | 1994-10-07 | 2005-12-20 | Saint Louis University | System for use in displaying images of a body part |
CA2201877C (en) | 1994-10-07 | 2004-06-08 | Richard D. Bucholz | Surgical navigation systems including reference and localization frames |
US5717414A (en) * | 1995-05-01 | 1998-02-10 | Lockheed-Martin Tactical Defense Systems | Video image tracking and mixing system |
US5592939A (en) | 1995-06-14 | 1997-01-14 | Martinelli; Michael A. | Method and system for navigating a catheter probe |
GB9515311D0 (en) * | 1995-07-26 | 1995-09-20 | 3D Scanners Ltd | Stripe scanners and methods of scanning |
WO1997011386A1 (en) * | 1995-09-21 | 1997-03-27 | Omniplanar, Inc. | Method and apparatus for determining position and orientation |
JP3796776B2 (en) * | 1995-09-28 | 2006-07-12 | ソニー株式会社 | Video / audio playback device |
US5828770A (en) * | 1996-02-20 | 1998-10-27 | Northern Digital Inc. | System for determining the spatial position and angular orientation of an object |
US6167145A (en) * | 1996-03-29 | 2000-12-26 | Surgical Navigation Technologies, Inc. | Bone navigation system |
US5832139A (en) * | 1996-07-31 | 1998-11-03 | Omniplanar, Inc. | Method and apparatus for determining degrees of freedom of a camera |
US6296613B1 (en) | 1997-08-22 | 2001-10-02 | Synthes (U.S.A.) | 3D ultrasound recording device |
US6226548B1 (en) | 1997-09-24 | 2001-05-01 | Surgical Navigation Technologies, Inc. | Percutaneous registration apparatus and method for use in computer-assisted surgical navigation |
US6021343A (en) | 1997-11-20 | 2000-02-01 | Surgical Navigation Technologies | Image guided awl/tap/screwdriver |
US6061644A (en) * | 1997-12-05 | 2000-05-09 | Northern Digital Incorporated | System for determining the spatial position and orientation of a body |
US6348058B1 (en) * | 1997-12-12 | 2002-02-19 | Surgical Navigation Technologies, Inc. | Image guided spinal surgery guide, system, and method for use thereof |
CA2335867C (en) | 1998-06-22 | 2008-12-30 | Synthes (U.S.A.) | Fiducial matching by means of fiducial screws |
US6477400B1 (en) | 1998-08-20 | 2002-11-05 | Sofamor Danek Holdings, Inc. | Fluoroscopic image guided orthopaedic surgery system with intraoperative registration |
US6388755B1 (en) | 1998-12-03 | 2002-05-14 | Advanced Optical Technologies, Inc. | Wireless position and orientation detecting system |
US6774885B1 (en) * | 1999-01-20 | 2004-08-10 | Motek B.V. | System for dynamic registration, evaluation, and correction of functional human behavior |
US6630915B1 (en) * | 1999-01-26 | 2003-10-07 | Lsa. Inc. | Wireless transmission system for transmitting data to a simulation system user |
CA2363138C (en) | 1999-03-01 | 2010-05-18 | Bae Systems Electronics Limited | Head tracker system |
GB9904669D0 (en) * | 1999-03-01 | 1999-04-21 | Marconi Electronic Syst Ltd | Head tracker system |
DE69931074T2 (en) | 1999-03-17 | 2006-11-16 | Synthes Ag Chur | DEVICE FOR PRESENTING AND PLANNING CRANE COATING OPERATIONS |
US6470207B1 (en) * | 1999-03-23 | 2002-10-22 | Surgical Navigation Technologies, Inc. | Navigational guidance via computer-assisted fluoroscopic imaging |
US6491699B1 (en) * | 1999-04-20 | 2002-12-10 | Surgical Navigation Technologies, Inc. | Instrument guidance method and system for image guided surgery |
AU766981B2 (en) | 1999-04-20 | 2003-10-30 | Ao Technology Ag | Device for the percutaneous obtainment of 3D-coordinates on the surface of a human or animal organ |
ATE242865T1 (en) * | 1999-05-03 | 2003-06-15 | Synthes Ag | POSITION DETECTION DEVICE WITH AIDS FOR DETERMINING THE DIRECTION OF THE GRAVITY VECTOR |
US6288704B1 (en) * | 1999-06-08 | 2001-09-11 | Vega, Vista, Inc. | Motion detection and tracking system to control navigation and display of object viewers |
US7180476B1 (en) * | 1999-06-30 | 2007-02-20 | The Boeing Company | Exterior aircraft vision system using a helmet-mounted display |
GB2352289B (en) * | 1999-07-14 | 2003-09-17 | Dennis Majoe | Position and orientation detection system |
GB9917591D0 (en) | 1999-07-28 | 1999-09-29 | Marconi Electronic Syst Ltd | Head tracker system |
US6424410B1 (en) | 1999-08-27 | 2002-07-23 | Maui Innovative Peripherals, Inc. | 3D navigation system using complementary head-mounted and stationary infrared beam detection units |
US7366562B2 (en) * | 2003-10-17 | 2008-04-29 | Medtronic Navigation, Inc. | Method and apparatus for surgical navigation |
US8239001B2 (en) * | 2003-10-17 | 2012-08-07 | Medtronic Navigation, Inc. | Method and apparatus for surgical navigation |
US6381485B1 (en) * | 1999-10-28 | 2002-04-30 | Surgical Navigation Technologies, Inc. | Registration of human anatomy integrated for electromagnetic localization |
US6747539B1 (en) | 1999-10-28 | 2004-06-08 | Michael A. Martinelli | Patient-shielding and coil system |
US6379302B1 (en) | 1999-10-28 | 2002-04-30 | Surgical Navigation Technologies Inc. | Navigation information overlay onto ultrasound imagery |
US8644907B2 (en) | 1999-10-28 | 2014-02-04 | Medtronic Navigaton, Inc. | Method and apparatus for surgical navigation |
US11331150B2 (en) | 1999-10-28 | 2022-05-17 | Medtronic Navigation, Inc. | Method and apparatus for surgical navigation |
US6493573B1 (en) | 1999-10-28 | 2002-12-10 | Winchester Development Associates | Method and system for navigating a catheter probe in the presence of field-influencing objects |
US6474341B1 (en) * | 1999-10-28 | 2002-11-05 | Surgical Navigation Technologies, Inc. | Surgical communication and power system |
US6499488B1 (en) | 1999-10-28 | 2002-12-31 | Winchester Development Associates | Surgical sensor |
WO2001031466A1 (en) | 1999-10-28 | 2001-05-03 | Winchester Development Associates | Coil structures and methods for generating magnetic fields |
AU3436001A (en) * | 1999-12-23 | 2001-07-03 | Justsystem Corporation | Method and apparatus for vision-based coupling between pointer actions and projected images |
US8788092B2 (en) | 2000-01-24 | 2014-07-22 | Irobot Corporation | Obstacle following sensor scheme for a mobile robot |
US8412377B2 (en) | 2000-01-24 | 2013-04-02 | Irobot Corporation | Obstacle following sensor scheme for a mobile robot |
US6725080B2 (en) | 2000-03-01 | 2004-04-20 | Surgical Navigation Technologies, Inc. | Multiple cannula image guided tool for image guided procedures |
US6956348B2 (en) | 2004-01-28 | 2005-10-18 | Irobot Corporation | Debris sensor for cleaning apparatus |
US6535756B1 (en) * | 2000-04-07 | 2003-03-18 | Surgical Navigation Technologies, Inc. | Trajectory storage apparatus and method for surgical navigation system |
JP4634570B2 (en) * | 2000-06-05 | 2011-02-16 | 株式会社東芝 | MRI equipment |
US7085400B1 (en) | 2000-06-14 | 2006-08-01 | Surgical Navigation Technologies, Inc. | System and method for image based sensor calibration |
US6795068B1 (en) * | 2000-07-21 | 2004-09-21 | Sony Computer Entertainment Inc. | Prop input device and method for mapping an object from a two-dimensional camera image to a three-dimensional space for controlling action in a game program |
DE10051415C2 (en) * | 2000-10-17 | 2003-10-09 | Advanced Realtime Tracking Gmb | Optical tracking system and method |
US7571511B2 (en) | 2002-01-03 | 2009-08-11 | Irobot Corporation | Autonomous floor-cleaning robot |
US6690134B1 (en) | 2001-01-24 | 2004-02-10 | Irobot Corporation | Method and system for robot localization and confinement |
US6636757B1 (en) * | 2001-06-04 | 2003-10-21 | Surgical Navigation Technologies, Inc. | Method and apparatus for electromagnetic navigation of a surgical probe near a metal object |
US7429843B2 (en) | 2001-06-12 | 2008-09-30 | Irobot Corporation | Method and system for multi-mode coverage for an autonomous robot |
US8396592B2 (en) * | 2001-06-12 | 2013-03-12 | Irobot Corporation | Method and system for multi-mode coverage for an autonomous robot |
US20030083844A1 (en) * | 2001-10-30 | 2003-05-01 | Reddi M. Mahadeva | Optical position sensing of multiple radiating sources in a movable body |
US9128486B2 (en) | 2002-01-24 | 2015-09-08 | Irobot Corporation | Navigational control system for a robotic device |
US6947786B2 (en) * | 2002-02-28 | 2005-09-20 | Surgical Navigation Technologies, Inc. | Method and apparatus for perspective inversion |
US6990368B2 (en) | 2002-04-04 | 2006-01-24 | Surgical Navigation Technologies, Inc. | Method and apparatus for virtual digital subtraction angiography |
US7998062B2 (en) | 2004-03-29 | 2011-08-16 | Superdimension, Ltd. | Endoscope structures and techniques for navigating to a target in branched structure |
US6892090B2 (en) * | 2002-08-19 | 2005-05-10 | Surgical Navigation Technologies, Inc. | Method and apparatus for virtual endoscopy |
US8386081B2 (en) | 2002-09-13 | 2013-02-26 | Irobot Corporation | Navigational control system for a robotic device |
US8428778B2 (en) | 2002-09-13 | 2013-04-23 | Irobot Corporation | Navigational control system for a robotic device |
US6828559B2 (en) * | 2002-11-14 | 2004-12-07 | Delphi Technologies, Inc | Sensor having a plurality of active areas |
US7697972B2 (en) * | 2002-11-19 | 2010-04-13 | Medtronic Navigation, Inc. | Navigation system for cardiac therapies |
US7599730B2 (en) * | 2002-11-19 | 2009-10-06 | Medtronic Navigation, Inc. | Navigation system for cardiac therapies |
US7542791B2 (en) * | 2003-01-30 | 2009-06-02 | Medtronic Navigation, Inc. | Method and apparatus for preplanning a surgical procedure |
US7660623B2 (en) | 2003-01-30 | 2010-02-09 | Medtronic Navigation, Inc. | Six degree of freedom alignment display for medical procedures |
US7063256B2 (en) * | 2003-03-04 | 2006-06-20 | United Parcel Service Of America | Item tracking and processing systems and methods |
WO2004079546A2 (en) | 2003-03-04 | 2004-09-16 | United Parcel Service Of America, Inc. | System for projecting a handling instruction onto a moving item or parcel |
US7570791B2 (en) * | 2003-04-25 | 2009-08-04 | Medtronic Navigation, Inc. | Method and apparatus for performing 2D to 3D registration |
US7313430B2 (en) * | 2003-08-28 | 2007-12-25 | Medtronic Navigation, Inc. | Method and apparatus for performing stereotactic surgery |
EP2316328B1 (en) | 2003-09-15 | 2012-05-09 | Super Dimension Ltd. | Wrap-around holding device for use with bronchoscopes |
ATE438335T1 (en) | 2003-09-15 | 2009-08-15 | Super Dimension Ltd | SYSTEM OF ACCESSORIES FOR USE WITH BRONCHOSCOPES |
US7835778B2 (en) | 2003-10-16 | 2010-11-16 | Medtronic Navigation, Inc. | Method and apparatus for surgical navigation of a multiple piece construct for implantation |
US7840253B2 (en) | 2003-10-17 | 2010-11-23 | Medtronic Navigation, Inc. | Method and apparatus for surgical navigation |
US8133115B2 (en) * | 2003-10-22 | 2012-03-13 | Sony Computer Entertainment America Llc | System and method for recording and displaying a graphical path in a video game |
US7332890B2 (en) | 2004-01-21 | 2008-02-19 | Irobot Corporation | Autonomous robot auto-docking and energy management systems and methods |
US7961909B2 (en) | 2006-03-08 | 2011-06-14 | Electronic Scripting Products, Inc. | Computer interface employing a manipulated object with absolute pose detection component and a display |
US8764725B2 (en) | 2004-02-09 | 2014-07-01 | Covidien Lp | Directional anchoring mechanism, method and applications thereof |
US7567834B2 (en) * | 2004-05-03 | 2009-07-28 | Medtronic Navigation, Inc. | Method and apparatus for implantation between two vertebral bodies |
US9008835B2 (en) | 2004-06-24 | 2015-04-14 | Irobot Corporation | Remote control scheduler and method for autonomous robotic device |
US8972052B2 (en) | 2004-07-07 | 2015-03-03 | Irobot Corporation | Celestial navigation system for an autonomous vehicle |
US7706917B1 (en) | 2004-07-07 | 2010-04-27 | Irobot Corporation | Celestial navigation system for an autonomous robot |
US7561717B2 (en) * | 2004-07-09 | 2009-07-14 | United Parcel Service Of America, Inc. | System and method for displaying item information |
US7289227B2 (en) * | 2004-10-01 | 2007-10-30 | Nomos Corporation | System and tracker for tracking an object, and related methods |
US20060071933A1 (en) | 2004-10-06 | 2006-04-06 | Sony Computer Entertainment Inc. | Application binary interface for multi-pass shaders |
US7403295B2 (en) * | 2004-10-25 | 2008-07-22 | Hoya Corporation | Position-detecting system |
US7636595B2 (en) * | 2004-10-28 | 2009-12-22 | Medtronic Navigation, Inc. | Method and apparatus for calibrating non-linear instruments |
FR2878615B1 (en) | 2004-11-30 | 2009-09-25 | Raquin Cyrille | SIMULATION SYSTEM FOR SHOOTING OR LAUNCHING PROJECTILE USING A SPECIFIC OBJECT OR LAUNCHER |
US7620476B2 (en) | 2005-02-18 | 2009-11-17 | Irobot Corporation | Autonomous surface cleaning robot for dry cleaning |
US8392021B2 (en) | 2005-02-18 | 2013-03-05 | Irobot Corporation | Autonomous surface cleaning robot for wet cleaning |
EP2289384B1 (en) * | 2005-02-18 | 2013-07-03 | iRobot Corporation | Autonomous surface cleaning robot for wet and dry cleaning |
US7590218B2 (en) * | 2005-03-23 | 2009-09-15 | Best Medical International, Inc. | System for monitoring the geometry of a radiation treatment apparatus, trackable assembly, program product, and related methods |
US8930023B2 (en) | 2009-11-06 | 2015-01-06 | Irobot Corporation | Localization by learning of wave-signal distributions |
US7636126B2 (en) * | 2005-06-22 | 2009-12-22 | Sony Computer Entertainment Inc. | Delay matching in audio/video systems |
US7835784B2 (en) | 2005-09-21 | 2010-11-16 | Medtronic Navigation, Inc. | Method and apparatus for positioning a reference frame |
JP4692250B2 (en) * | 2005-12-01 | 2011-06-01 | 株式会社島津製作所 | Head motion tracker |
US7441298B2 (en) | 2005-12-02 | 2008-10-28 | Irobot Corporation | Coverage robot mobility |
ES2522926T3 (en) | 2005-12-02 | 2014-11-19 | Irobot Corporation | Autonomous Cover Robot |
ES2334064T3 (en) * | 2005-12-02 | 2010-03-04 | Irobot Corporation | MODULAR ROBOT. |
KR101099808B1 (en) | 2005-12-02 | 2011-12-27 | 아이로보트 코퍼레이션 | Robot system |
EP2816434A3 (en) * | 2005-12-02 | 2015-01-28 | iRobot Corporation | Autonomous coverage robot |
US9168102B2 (en) * | 2006-01-18 | 2015-10-27 | Medtronic Navigation, Inc. | Method and apparatus for providing a container to a sterile environment |
US20070239169A1 (en) * | 2006-03-17 | 2007-10-11 | Perception Raisonnement Action En Medecine | Reference marker and use in a motion tracking system |
US8112292B2 (en) | 2006-04-21 | 2012-02-07 | Medtronic Navigation, Inc. | Method and apparatus for optimizing a therapy |
US7880746B2 (en) * | 2006-05-04 | 2011-02-01 | Sony Computer Entertainment Inc. | Bandwidth management through lighting control of a user environment via a display device |
US7965859B2 (en) | 2006-05-04 | 2011-06-21 | Sony Computer Entertainment Inc. | Lighting control of a user environment via a display device |
US20090044370A1 (en) | 2006-05-19 | 2009-02-19 | Irobot Corporation | Removing debris from cleaning robots |
US8417383B2 (en) | 2006-05-31 | 2013-04-09 | Irobot Corporation | Detecting robot stasis |
US8660635B2 (en) | 2006-09-29 | 2014-02-25 | Medtronic, Inc. | Method and apparatus for optimizing a computer assisted surgical procedure |
DE102007021185B4 (en) | 2007-05-05 | 2012-09-20 | Ziehm Imaging Gmbh | X-ray diagnostic device with a plurality of coded marks and a method for determining the position of device parts of the X-ray diagnostic device |
KR101160393B1 (en) | 2007-05-09 | 2012-06-26 | 아이로보트 코퍼레이션 | Compact autonomous coverage robot |
JP5027735B2 (en) * | 2007-05-25 | 2012-09-19 | サッポロビール株式会社 | Method for producing sparkling alcoholic beverage |
ES2452037T3 (en) | 2007-09-26 | 2014-03-31 | Elbit Systems Ltd. | Wide field of vision optical tracking system |
US8885177B2 (en) | 2007-09-26 | 2014-11-11 | Elbit Systems Ltd. | Medical wide field of view optical tracking system |
US8905920B2 (en) | 2007-09-27 | 2014-12-09 | Covidien Lp | Bronchoscope adapter and method |
US8014003B2 (en) * | 2008-01-26 | 2011-09-06 | Avigdor Ronn | Body metric differential measurement device |
US9575140B2 (en) | 2008-04-03 | 2017-02-21 | Covidien Lp | Magnetic interference detection system and method |
US8473032B2 (en) | 2008-06-03 | 2013-06-25 | Superdimension, Ltd. | Feature-based registration method |
US8218847B2 (en) | 2008-06-06 | 2012-07-10 | Superdimension, Ltd. | Hybrid registration method |
US8932207B2 (en) | 2008-07-10 | 2015-01-13 | Covidien Lp | Integrated multi-functional endoscopic tool |
US8165658B2 (en) | 2008-09-26 | 2012-04-24 | Medtronic, Inc. | Method and apparatus for positioning a guide relative to a base |
US8422777B2 (en) * | 2008-10-14 | 2013-04-16 | Joshua Victor Aller | Target and method of detecting, identifying, and determining 3-D pose of the target |
PT104224B (en) * | 2008-10-21 | 2010-12-27 | Univ Do Porto | SYSTEM OF MOTION CAPTURE AND LOCATION OF OBJECTS TRAVELING TO COLORED AND PULSED LIGHT SENSORS WITH STEREOSCOPIC VISION SYSTEM |
US8175681B2 (en) | 2008-12-16 | 2012-05-08 | Medtronic Navigation Inc. | Combination of electromagnetic and electropotential localization |
US8611984B2 (en) | 2009-04-08 | 2013-12-17 | Covidien Lp | Locatable catheter |
DE102009016498A1 (en) * | 2009-04-08 | 2010-10-21 | Ventech Gmbh | Method and device for determining the tread depth of a vehicle tire |
US8494613B2 (en) | 2009-08-31 | 2013-07-23 | Medtronic, Inc. | Combination localization system |
US8494614B2 (en) * | 2009-08-31 | 2013-07-23 | Regents Of The University Of Minnesota | Combination localization system |
US9189885B2 (en) | 2009-09-16 | 2015-11-17 | Knorr-Bremse Systeme Fur Schienenfahrzeuge Gmbh | Visual presentation system |
CN104127156B (en) | 2010-02-16 | 2017-01-11 | 艾罗伯特公司 | Vacuum Brush |
US10786736B2 (en) | 2010-05-11 | 2020-09-29 | Sony Interactive Entertainment LLC | Placement of user information in a game space |
US10582834B2 (en) | 2010-06-15 | 2020-03-10 | Covidien Lp | Locatable expandable working channel and method |
DE102011009463A1 (en) * | 2011-01-24 | 2012-07-26 | H&D Systems GmbH | Device for determining position of objects e.g. tools, in manufacturing industry, has computing unit computing exact position of light source using data that is received from camera or light sensor, where source is arranged on object |
US9342817B2 (en) | 2011-07-07 | 2016-05-17 | Sony Interactive Entertainment LLC | Auto-creating groups for sharing photos |
US9747680B2 (en) | 2013-11-27 | 2017-08-29 | Industrial Technology Research Institute | Inspection apparatus, method, and computer program product for machine vision inspection |
WO2015131171A1 (en) | 2014-02-28 | 2015-09-03 | Decision Sciences International Corporation | Charged particle tomography for anatomical imaging |
US10555709B2 (en) | 2014-02-28 | 2020-02-11 | Decision Sciences International Corporation | Charged particle tomography scanner for real-time volumetric radiation dose monitoring and control |
US10952593B2 (en) | 2014-06-10 | 2021-03-23 | Covidien Lp | Bronchoscope adapter |
EP4194880A1 (en) | 2015-03-05 | 2023-06-14 | Atracsys Sàrl | Redundant reciprocal tracking system |
US10426555B2 (en) | 2015-06-03 | 2019-10-01 | Covidien Lp | Medical instrument with sensor for use in a system and method for electromagnetic navigation |
US9962134B2 (en) | 2015-10-28 | 2018-05-08 | Medtronic Navigation, Inc. | Apparatus and method for maintaining image quality while minimizing X-ray dosage of a patient |
US10156912B2 (en) * | 2015-11-25 | 2018-12-18 | Honeywell International Inc. | High speed, high precision six degree-of-freedom optical tracker system and method |
US10303246B2 (en) * | 2016-01-20 | 2019-05-28 | North Inc. | Systems, devices, and methods for proximity-based eye tracking |
US10478254B2 (en) | 2016-05-16 | 2019-11-19 | Covidien Lp | System and method to access lung tissue |
US11577159B2 (en) | 2016-05-26 | 2023-02-14 | Electronic Scripting Products Inc. | Realistic virtual/augmented/mixed reality viewing and interactions |
CN106405498B (en) * | 2016-08-31 | 2019-08-23 | 珠海真幻科技有限公司 | The method and apparatus of mobile device space location is realized with multiple light courcess |
US10418705B2 (en) | 2016-10-28 | 2019-09-17 | Covidien Lp | Electromagnetic navigation antenna assembly and electromagnetic navigation system including the same |
US10751126B2 (en) | 2016-10-28 | 2020-08-25 | Covidien Lp | System and method for generating a map for electromagnetic navigation |
US10722311B2 (en) | 2016-10-28 | 2020-07-28 | Covidien Lp | System and method for identifying a location and/or an orientation of an electromagnetic sensor based on a map |
US10446931B2 (en) | 2016-10-28 | 2019-10-15 | Covidien Lp | Electromagnetic navigation antenna assembly and electromagnetic navigation system including the same |
US10517505B2 (en) | 2016-10-28 | 2019-12-31 | Covidien Lp | Systems, methods, and computer-readable media for optimizing an electromagnetic navigation system |
US10638952B2 (en) | 2016-10-28 | 2020-05-05 | Covidien Lp | Methods, systems, and computer-readable media for calibrating an electromagnetic navigation system |
US10792106B2 (en) | 2016-10-28 | 2020-10-06 | Covidien Lp | System for calibrating an electromagnetic navigation system |
US10615500B2 (en) | 2016-10-28 | 2020-04-07 | Covidien Lp | System and method for designing electromagnetic navigation antenna assemblies |
US10580386B2 (en) * | 2017-04-21 | 2020-03-03 | Ford Global Technologies, Llc | In-vehicle projected reality motion correction |
US10471478B2 (en) | 2017-04-28 | 2019-11-12 | United Parcel Service Of America, Inc. | Conveyor belt assembly for identifying an asset sort location and methods of utilizing the same |
US11219489B2 (en) | 2017-10-31 | 2022-01-11 | Covidien Lp | Devices and systems for providing sensors in parallel with medical tools |
CN108519054B (en) * | 2018-04-24 | 2019-12-17 | 长春理工大学 | calibration device and calibration method for arc-shaped infrared target simulator |
Family Cites Families (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3032660A (en) * | 1959-12-17 | 1962-05-01 | Fairchild Camera Instr Co | Range finder and tracking device |
US3375375A (en) * | 1965-01-08 | 1968-03-26 | Honeywell Inc | Orientation sensing means comprising photodetectors and projected fans of light |
US3917412A (en) * | 1972-04-11 | 1975-11-04 | Us Navy | Advanced helmet tracker using lateral photodetection and light-emitting diodes |
US3916094A (en) * | 1974-06-21 | 1975-10-28 | Us Navy | Submersible visual simulator for remotely piloted systems |
US3951550A (en) * | 1974-08-12 | 1976-04-20 | The Magnavox Company | Direction-sensing virtual aperture radiation detector |
GB1489758A (en) * | 1974-10-16 | 1977-10-26 | Redifon Ltd | Visual display apparatus |
GB1550339A (en) * | 1975-08-28 | 1979-08-15 | Elliott Brothers London Ltd | Optical angular position sensors |
GB1520154A (en) * | 1976-02-24 | 1978-08-02 | Elliott Brothers London Ltd | Apparatus for measuring the angular displacement of a bod |
US4028725A (en) * | 1976-04-21 | 1977-06-07 | Grumman Aerospace Corporation | High-resolution vision system |
FR2399033A1 (en) * | 1977-07-29 | 1979-02-23 | Thomson Csf | DEVICE FOR LOCATING A RADIANT SOURCE AND DIRECTION TRACKING SYSTEM INCLUDING SUCH A DEVICE |
FR2416480A1 (en) * | 1978-02-03 | 1979-08-31 | Thomson Csf | RADIANT SOURCE LOCATION DEVICE AND STEERING TRACKING SYSTEM INCLUDING SUCH A DEVICE |
GB2043938B (en) * | 1978-12-21 | 1983-08-17 | Redifon Simulation Ltd | Visual display apparatus |
GB2115179B (en) * | 1978-12-21 | 1984-01-18 | Redifon Simulation Ltd | Improvements in or relating to visual display apparatus |
GB2043290B (en) * | 1979-01-11 | 1983-08-17 | Redifon Simulation Ltd | Visual display apparatus |
GB2043939A (en) * | 1979-01-11 | 1980-10-08 | Redifon Simulation Ltd | Visual display apparatus |
GB2043289A (en) * | 1979-01-11 | 1980-10-01 | Redifon Simulation Ltd | Improvements in or relating to visual display apparatus |
GB2041563B (en) * | 1979-01-11 | 1983-08-17 | Redifon Simulation Ltd | Visual display apparatus |
FR2451063A1 (en) * | 1979-03-09 | 1980-10-03 | Crouzet Sa | DEVICE FOR DETERMINING THE DIRECTION OF A SIGHTLINE |
US4209255A (en) * | 1979-03-30 | 1980-06-24 | United Technologies Corporation | Single source aiming point locator |
US4419012A (en) * | 1979-09-11 | 1983-12-06 | Elliott Brothers (London) Limited | Position measuring system |
US4348186A (en) * | 1979-12-17 | 1982-09-07 | The United States Of America As Represented By The Secretary Of The Navy | Pilot helmet mounted CIG display with eye coupled area of interest |
US4348185A (en) * | 1980-02-14 | 1982-09-07 | The United States Of America As Represented By The Secretary Of The Navy | Wide angle infinity display system |
FR2487077A1 (en) * | 1980-07-18 | 1982-01-22 | Trt Telecom Radio Electr | DEVICE FOR REMOTELY DETERMINING THE POSITION IN THE SPACE OF AN OBJECT PERFORMING ROTATION MOVEMENTS |
US4446480A (en) * | 1981-12-14 | 1984-05-01 | The United States Of America As Represented By The Secretary Of The Navy | Head position and orientation sensor |
SE430539B (en) * | 1982-03-24 | 1983-11-21 | Asea Ab | FIBEROPTIC METHOD SYSTEM FOR METANDE, HANDLING AND PROCESSING BODIES |
-
1984
- 1984-05-22 US US06/613,063 patent/US4649504A/en not_active Expired - Fee Related
- 1984-07-18 CA CA000459133A patent/CA1206617A/en not_active Expired
-
1985
- 1985-05-22 DE DE8585303604T patent/DE3579363D1/en not_active Expired - Lifetime
- 1985-05-22 EP EP85303604A patent/EP0162713B1/en not_active Expired - Lifetime
- 1985-05-22 JP JP10845885A patent/JPS6134406A/en active Pending
Also Published As
Publication number | Publication date |
---|---|
US4649504A (en) | 1987-03-10 |
JPS6134406A (en) | 1986-02-18 |
EP0162713A2 (en) | 1985-11-27 |
EP0162713B1 (en) | 1990-08-29 |
DE3579363D1 (en) | 1990-10-04 |
EP0162713A3 (en) | 1987-03-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CA1206617A (en) | Optical position and orientation measurement techniques | |
Bhatnagar | Position trackers for Head Mounted Display systems: A survey | |
Junkins et al. | Vision-based navigation for rendezvous, docking and proximity operations | |
US4956794A (en) | Single camera three dimensional head position sensing system | |
EP0482330A2 (en) | Optical position and orientation sensor | |
EP0181553A1 (en) | 3-D active vision sensor | |
DE3886267T2 (en) | Arrangement for measuring an angular displacement of an object. | |
US4923303A (en) | Measuring device for location and attitude determination of an object | |
CN107909624B (en) | Method for extracting and fusing two-dimensional image from three-dimensional tomography | |
ATE152248T1 (en) | PRECISION LANDING SYSTEM | |
US20100098325A1 (en) | System for optically detecting position and/or orientation of objects comprising at least two coplanar sensors | |
Schaefer et al. | Accuracy of position detection using a position-sensitive detector | |
GB2157426A (en) | Angular position sensors | |
US20100020335A1 (en) | Method for detecting the space orientation and position of an object | |
Frey et al. | Off-the-shelf, real-time, human body motion capture for synthetic environments | |
Blate et al. | Implementation and Evaluation of a 50 kHz, $28\mu\mathrm {s} $ Motion-to-Pose Latency Head Tracking Instrument | |
Madritsch et al. | CCD‐Camera Based Optical Beacon Tracking for Virtual and Augmented Reality | |
Blais et al. | A very compact real time 3-D range sensor for mobile robot applications | |
GB2090657A (en) | Photoelectric Method and Apparatus for Detecting the Position of Edge-defined Structures | |
US4045140A (en) | Means for near real time C-W laser source characterization | |
GB2301966A (en) | System for determining the orientation and location of a moving body with respect to a structure | |
KR100269180B1 (en) | Radar device | |
CN213658955U (en) | Portable device for visually positioning noise source | |
Feron et al. | A passive sensor for position and attitude estimation using an interferometric target | |
JPS6175265A (en) | 3-d speedmeter |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
MKEX | Expiry |