CN102222348A - Method for calculating three-dimensional object motion vector - Google Patents

Method for calculating three-dimensional object motion vector Download PDF

Info

Publication number
CN102222348A
CN102222348A CN 201110176736 CN201110176736A CN102222348A CN 102222348 A CN102222348 A CN 102222348A CN 201110176736 CN201110176736 CN 201110176736 CN 201110176736 A CN201110176736 A CN 201110176736A CN 102222348 A CN102222348 A CN 102222348A
Authority
CN
China
Prior art keywords
dimensional
motion vector
video camera
prime
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN 201110176736
Other languages
Chinese (zh)
Other versions
CN102222348B (en
Inventor
袁杰
顾人舒
石磊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University
Original Assignee
Nanjing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University filed Critical Nanjing University
Priority to CN 201110176736 priority Critical patent/CN102222348B/en
Publication of CN102222348A publication Critical patent/CN102222348A/en
Application granted granted Critical
Publication of CN102222348B publication Critical patent/CN102222348B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a method for calculating a three-dimensional object motion vector. The method comprises the following steps of: detection and matching of object corner: identifying objects in a video steam, labelling the corners of the objects in a basic frame and a key frame, and matching the corners with each other; calculating the three-dimensional motion vector of the key frame in relation to the basic frame based on the matched corners in the basic frame and the key frame; calculating a three-dimensional object model of the key frame based on a basic three-dimensional model and the three-dimensional motion vector, back projecting the three-dimensional object model to a two-dimensional plane, eliminating wrong points and calculating the final three-dimensional object motion vector.

Description

A kind of objective motion vector calculating method
Technical field
The present invention relates to a kind of multi-angle dynamic imaging field of three-dimensional body, the target three-dimensional motion vector computing method during particularly a kind of stereoscopic video with free visual angles dynamically shows.
Background technology
Along with technical development, occurred some in the research field and shown the display terminal of stereoeffect, wherein, free-viewing angle stereo display is active, real dynamic the demonstration, do not rely on parallax and form the stereoscopic vision effect, the observer can independently select observation visual angle and distance.The calculating of target three-dimensional motion vector is a difficult point in the free-viewing angle stereo display, also be stereoscopic vision from static state to dynamic key.By calculating the Three-dimension Target motion vector, can follow the tracks of, the motion of three-dimensional body in the simulation, display video, therefore for video signal processing field, important function and significance is arranged, be necessary procedure crucial in the Stereoscopic Video Presentation.
Summary of the invention
Goal of the invention: technical matters to be solved by this invention is at the Stereoscopic Video Presentation system, and a kind of linear method that calculates the target three-dimensional motion vector is provided.
Technical scheme: the invention discloses the calculating of three-dimensional motion vector in a kind of Stereoscopic Video Presentation, may further comprise the steps:
Step (1), target Corner Detection and coupling: the target in the identification video stream marks out the angle point of target in basic frame and the key frame, and mates;
Step (2), according to the angle point that mates in basic frame and the key frame, the primary Calculation key frame is with respect to the three-dimensional motion of basic frame;
Step (3), according to the target three-dimensional model of motion and basic three-dimensional model calculating key frame, it is overdue that the two dimensional surface rejecting is returned by back projection, calculates final three-dimensional motion vector.
Among the present invention, step (1) may further comprise the steps:
Step (11), the target in identification key frame and the basic frame;
Step (12) marks out the angle point of target in basic frame and the key frame;
Step (13) is mated key frame and basic frame: the corresponding angle point and the two-dimensional coordinate thereof that find key frame and basic frame.
Among the present invention, step (2) may further comprise the steps:
Based on video camera 1 and video camera 2:
Use approximation method to draw system of linear equations:
x 1 ′ - p 14 - X 1 p 11 - Y 1 p 12 - Z 1 p 13 y 1 ′ - p 24 - X 1 p 21 - Y 1 p 22 - Z 1 p 23 M M x n ′ - pp 14 - X n pp 11 - Y n pp 12 - Z n pp 13 y n ′ - pp 24 - X n pp 21 - Y n pp 22 - Z n pp 23 = Y 1 p 13 - Z 1 p 12 Z 1 p 11 - X 1 p 13 X 1 p 12 - Y 1 p 11 p 11 p 12 p 13 Y 1 p 23 - Z 1 p 22 Z 1 p 21 - X 1 p 23 X 1 p 22 - Y 1 p 21 p 21 p 22 p 23 M M Y n pp 13 - Z n pp 12 Z n pp 11 - X n pp 13 X n pp 12 - Y n pp 11 pp 11 pp 12 pp 13 Y n pp 23 - Z n pp 22 Z n pp 21 - X n pp 23 X n pp 22 - Y n pp 21 pp 21 pp 22 pp 23 φ 1 φ 2 φ 3 t 1 t 2 t 3 ,
φ 1, φ 2, φ 3Be respectively the three-dimensional motion vector reduced parameter, t 1, t 2, t 3Be respectively the translation matrix parameter; Form three-dimensional motion vector M:
M = 1 - φ 3 φ 2 t 1 φ 3 1 - φ 1 t 2 - φ 2 φ 1 1 t 3 0 0 0 1 ;
Or use non-approximation method, draw system of linear equations:
x 1 ′ - p 14 y 1 ′ - p 24 M M M M M M M x n ′ - pp 14 y n ′ - pp 24 = X 1 p 11 X 1 p 12 X 1 p 13 Y 1 p 11 Y 1 p 12 Y 1 p 13 Z 1 p 11 Z 1 p 12 Z 1 p 13 p 11 p 12 p 13 X 1 p 21 X 1 p 22 X 1 p 23 Y 1 p 21 Y 1 p 22 Y 1 p 23 Z 1 p 21 Z 1 p 22 Z 1 p 23 p 21 p 22 p 23 . M M M M M M X n pp 11 X n pp 12 X n pp 13 Y n pp 11 Y n pp 12 Y n pp 13 Z n pp 11 Z n pp 12 Z n pp 13 pp 11 pp 12 pp 13 X n pp 21 X n pp 22 X n pp 23 Y n pp 21 Y n pp 22 Y n pp 23 Z n pp 21 Z n pp 22 Z n pp 23 pp 21 pp 22 pp 23 r 11 r 21 r 31 r 12 r 22 r 32 r 13 r 23 r 33 t 1 t 2 t 3
, r wherein 11, r 21, r 31, r 12, r 22, r 32, r 13, r 23, r 33Be the rotation matrix parameter, form three-dimensional motion vector M:
M = r 11 r 12 r 13 t 1 r 21 r 22 r 23 t 2 r 31 r 32 r 33 t 3 0 0 0 1 ;
(X wherein i, Y i, Z i) any some three-dimensional coordinate in the basic frame of video camera 1 or video camera 2 of expression;
(X i', Y i', Z i') represent this three-dimensional coordinate in the key frame of video camera 1 or video camera 2; (x i, y i) represent that this projects to the two-dimensional coordinate of two dimensional surface at video camera 1 or video camera 2 in basic frame; (x i', y i') represent that this projects to the two-dimensional coordinate of two dimensional surface at video camera 1 or video camera 2 in key frame;
P = p 11 p 12 p 13 p 14 p 21 p 22 p 23 p 24 0 0 0 1 Projection matrix for video camera 1;
PP = pp 11 pp 12 pp 13 pp 14 pp 21 pp 22 pp 23 pp 24 0 0 0 1 Projection matrix for video camera 2;
Find the solution the least mean-square error of system of linear equations and separate, obtain the kinematic matrix parameter, i.e. motion vector.
Among the present invention, step (3) may further comprise the steps:
To two video cameras, described kinematic matrix parameter is additional to basic three-dimensional model, obtain the three-dimensional model of two camera video key frames;
To two video cameras, respectively with the three-dimensional model of two camera video key frames, the two dimensional surface of video camera matrix correspondence separately returns in back projection, with the angle point two-dimensional coordinate of the key frame of corresponding video camera relatively, reject the point of error beyond threshold value, use step 2 to recomputate the kinematic matrix parameter, obtain final three-dimensional motion motion vector.
Beneficial effect: the present invention calculates relative motion fast and obtains the key frame three-dimensional model key frame according to the static three-dimensional model of basic frame, so that obtain the motion video at any visual angle of Same Scene.The present invention does not need camera calibration, and real-time is good, and the smoothness that helps dynamic video shows, and algorithm has certain robustness.In calculating the free-viewing angle display system, can carry out three-dimensional reconstruction consuming time every the long period, and calculate motion in real time with the inventive method, adjust the position of three-dimensional body, make to show that real-time is better, do not sacrifice precision simultaneously.
Description of drawings
Below in conjunction with the drawings and specific embodiments the present invention is done further to specify, above-mentioned and/or otherwise advantage of the present invention will become apparent.
Fig. 1 is video sequence basis frame of the present invention and key frame signal.
The coordinate system perspective view that Fig. 2 is involved in the present invention.
Fig. 3 is the control experiment comparing result of the present invention with existing research.
Embodiment:
The inventive method adopts twin camera (can expand to multiple-camera) on the basis of static three-dimensional reconstruction, according to two angle motion video pictures of synchronous shooting Same Scene, calculate the target three-dimensional motion vector, shows so that realize the dynamic solid video.
During the invention process, the condition that needs to satisfy is:
Video flowing is synchronous: the collection that starts video flowing comes the two-dimensional scene of synchronous different visual angles camera head shooting; Concerning embodiment, be embodied as the picture frame that adopts with the moment two video cameras, as Fig. 1.
During the invention process, Given information has:
Basis frame and basic three-dimensional model: basic three-dimensional model is that the center is set up with basic frame, and basic frame comprises the basic frame of two video cameras.
The projection matrix of two video cameras: rebuild in the process of basic three-dimensional model and can obtain the video camera projection matrix.
Among the present invention, on the completed basis of static three-dimensional reconstruction, the two-dimensional video picture according to two video cameras provide calculates the Three-dimension Target motion vector.As shown in Figure 1, the image sequence that the optional interval of key frame equates, present embodiment choose wherein that a frame calculates its motion vector with respect to basic frame, and other frame method is identical.
Implementation process is as follows:
One, target Corner Detection and coupling: discern two video cameras target in the video flowing separately, mark out in two basic separately frames of video camera and the key frame angle point on the target;
Can be 200910234584.8 in the application number of application on November 23rd, 2009 referring to the applicant, name be called the patent of invention of " a kind of method for displaying stereoscopic video with free visual angles ".Concise and to the point step is:
1.1 with the Harris algorithm angle point in the basic frame being carried out the first time estimates;
1.2 the angle point that estimates is further screened with the SUSAN algorithm, draws final angle point.
1.3 key frame and basic frame are mated: the two-dimensional coordinate (x that finds key frame with the template window matching process 1, y 1) and the two-dimensional coordinate (x of the corresponding angle point of basic frame 1', y 1').
Use above method, for two video cameras, its key frame is mated to basic frame respectively.
Two, calculate the three-dimensional motion of key frame with respect to basic frame.
The present invention adopts the affine camera model under the homogeneous coordinates, and is suitable when its degree of depth can be ignored relatively in the change in depth of target as shown in Figure 2, realistic video camera.
Consider video camera 1 earlier.For more arbitrarily, establish that three-dimensional coordinate is respectively (X in basic frame and the key frame 1, Y 1, Z 1) and (X 1', Y 1', Z 1'), two-dimensional coordinate is respectively (x 1, y 1) and (x 1', y 1').
Three-dimensional coordinate is write to the mapping of two-dimensional coordinate:
x 1 y 1 1 = P X 1 Y 1 Z 1 1 - - - ( 1 )
Wherein P = p 11 p 12 p 13 p 14 p 21 p 22 p 23 p 24 0 0 0 1 Be the video camera projection matrix.
Then key frame can be expressed as with respect to the motion of basic frame:
X 1 ′ Y 1 ′ Z 1 ′ 1 = M X 1 Y 1 Z 1 1 - - - ( 2 )
Wherein kinematic matrix M is: M = R T O 1 = r 11 r 12 r 13 t 1 r 21 r 22 r 23 t 2 r 31 r 32 r 33 t 3 0 0 0 1 - - - ( 3 )
Wherein R is a rotation matrix, and T is a translation matrix, and all elements of 1 * 3 matrix O is zero.Though rotation matrix R has 9 elements, it is an orthogonal matrix, has 6 independent restrainings:
Figure BDA0000071689480000056
Thereby real independent parameter has only 3.Can be by Roll-Pitch-Yaw (x of right-handed Cartesian coordinate system, y and z axle) method representation rotation matrix, θ is the anglec of rotation:
R = cos θ y cos θ z - cos θ y sin θ z sin θ y sin θ x sin θ y cos θ z + cos θ x sin θ z - sin θ x sin θ y sin θ z + cos θ x cos θ z - sin θ x cos θ y - cos θ x sin θ y cos θ z + sin θ x sin θ z cos θ x sin θ y sin θ z + sin θ x cos θ z cos θ x cos θ y - - - ( 4 )
Finding the solution of motion is divided into 2 steps:
2.1 will finding the solution motion, step 1. is summed up as the group b=Ax that solves an equation
By x 1 y 1 1 = P X 1 Y 1 Z 1 1 x 1 ′ y 1 ′ 1 = P X 1 ′ Y 1 ′ Z 1 ′ 1 X 1 ′ Y 1 ′ Z 1 ′ 1 = M X 1 Y 1 Z 1 1 Three formulas can get:
x 1 ′ y 1 ′ 1 = PM X 1 Y 1 Z 1 1 - - - ( 5 )
Each point provides two equations, and is as follows:
x 1 ′ = p 14 + p 11 t 1 + p 12 t 2 + p 13 t 3 + X 1 ( p 11 r 11 + p 12 r 21 + p 13 r 31 ) + Y 1 ( p 11 r 12 + p 12 r 22 + p 13 r 32 ) + Z 1 ( p 11 r 13 + p 12 r 23 + p 13 r 33 ) y 1 ′ = p 24 + p 21 t 1 + p 22 t 2 + p 23 t 3 + X 1 ( p 21 r 11 + p 22 r 21 + p 23 r 31 ) + Y 1 ( p 21 r 12 + p 22 r 22 + p 23 r 32 ) + Z 1 ( p 21 r 13 + p 22 r 23 + p 23 r 33 ) - - - ( 6 )
1. " if make small angle approximation (key frame is separated by when nearer, and it is less move, can make small angle approximation), and the Roll of employing (6) formula, Pitch and Yaw method for expressing, rotation matrix is reduced to, φ 1, φ 2, φ 3Be respectively the three-dimensional motion vector reduced parameter:
R = 1 - n 3 θ n 2 θ n 3 θ 1 - n 1 θ - n 2 θ n 1 θ 1 = 1 - φ 3 φ 2 φ 3 1 - φ 1 - φ 2 φ 1 1 - - - ( 7 )
(n wherein 1θ) 2+ (n 2θ) 2+ (n 3θ) 22So, φ 1 2+ φ 2 2+ φ 3 22(8)
System of equations (6) is rewritten as:
x 1 ′ - p 14 - X 1 p 11 - Y 1 p 12 - Z 1 p 13 y 1 ′ - p 24 - X 1 p 21 - Y 1 p 22 - Z 1 p 23 = Y 1 p 13 - Z 1 p 12 Z 1 p 11 - X 1 p 13 X 1 p 12 - Y 1 p 11 p 11 p 12 p 13 Y 1 p 23 - Z 1 p 22 Z 1 p 21 - X 1 p 23 X 1 p 22 - Y 1 p 21 p 21 p 22 p 23 φ 1 φ 2 φ 3 t 1 t 2 t 3 - - - ( 9 )
As only using the multiple spot of a video camera, can list system of equations:
x 1 ′ - p 14 - X 1 p 11 - Y 1 p 12 - Z 1 p 13 y 1 ′ - p 24 - X 1 p 21 - Y 1 p 22 - Z 1 p 23 . . x n ′ - p 14 - X n p 11 - Y n p 12 - Z n p 13 y n ′ - p 24 - X n p 21 - Y n p 22 - Z n p 23 = Y 1 p 13 - Z 1 p 12 Z 1 p 11 - X 1 p 13 X 1 p 12 - Y 1 p 11 p 11 p 12 p 13 Y 1 p 23 - Z 1 p 22 Z 1 p 21 - X 1 p 23 X 1 p 22 - Y 1 p 21 p 21 p 22 p 23 . . Y n p 13 - Z n p 12 Z n p 11 - X n p 13 X n p 12 - Y n p 11 p 11 p 12 p 13 Y n p 23 - Z n p 22 Z n p 21 - X n p 23 X n p 22 - Y n p 21 p 21 p 22 p 23 φ 1 φ 2 φ 3 t 1 t 2 t 3 - - - ( 10 )
First matrix of equation the right is designated as A in the following formula, and its order is 5 to the maximum.This is because a video camera can not provide along the depth information of its optical axis direction.Issued a certificate below:
I. observation matrix A finds that right three row interlacing are identical.This matrix is done line translation, with the first row cancellation 2i-1 capable after three row, with the second row cancellation 2i capable after three row (i=2,3 ... n):
Y 1 p 13 - Z 1 p 12 Z 1 p 11 - X 1 p 13 X 1 p 12 - Y 1 p 1 p 11 p 12 p 13 Y 1 p 23 - Z 1 p 22 Z 1 p 21 - X 1 p 23 X 1 p 22 - Y 1 p 21 p 21 p 22 p 23 M 0 0 0 M M M 0 0 0 Y n p 13 - Z n p 12 Z n p 11 - X n p 13 X n p 12 - Y n p 11 0 0 0 Y n p 23 - Z n p 22 Z n p 21 - X n p 23 X n p 22 - Y n p 21 0 0 0 - - - ( 11 )
Ii. in the matrix that obtains, right three column ranks are 2 to the maximum because non-vanishing have only two the row; A left side three column ranks are 3 to the maximum.So this rank of matrix is 5 to the maximum.The elementary row rank transformation does not change rank of matrix, so the order of matrix A is 5 to the maximum.
By above-mentioned proof as can be known, find the solution the rigid body three-dimensional motion, need two video cameras at least.Second video camera projection matrix is:
PP = pp 11 pp 12 pp 13 pp 14 pp 21 pp 22 pp 23 pp 24 0 0 0 1
So will find the solution motion, the group that need establish an equation to the multiple spot of two video cameras is:
x 1 ′ - p 14 - X 1 p 11 - Y 1 p 12 - Z 1 p 13 y 1 ′ - p 24 - X 1 p 21 - Y 1 p 22 - Z 1 p 23 M M x n ′ - pp 14 - X n pp 11 - Y n pp 12 - Z n pp 13 y n ′ - pp 24 - X n pp 21 - Y n pp 22 - Z n pp 23 = Y 1 p 13 - Z 1 p 12 Z 1 p 11 - X 1 p 13 X 1 p 12 - Y 1 p 11 p 11 p 12 p 13 Y 1 p 23 - Z 1 p 22 Z 1 p 21 - X 1 p 23 X 1 p 22 - Y 1 p 21 p 21 p 22 p 23 M M Y n pp 13 - Z n pp 12 Z n pp 11 - X n pp 13 X n pp 12 - Y n pp 11 pp 11 pp 12 pp 13 Y n pp 23 - Z n pp 22 Z n pp 21 - X n pp 23 X n pp 22 - Y n pp 21 pp 21 pp 22 pp 23 φ 1 φ 2 φ 3 t 1 t 2 t 3 - - - ( 12 )
Note equation left side matrix is b, and the right matrix is followed successively by A and x.This moment, the order of A was 6 to the maximum.. then finding the solution conversion of motion is solving equation group b=Ax.
N point comprises m point of video camera 1, and the k of video camera 2 point.N=m+k (m, k 〉=1). in theory, 3 groups of points that n=3 promptly uses two video cameras to the time, can solve system of equations; When n 〉=3, need separate the overdetermined equation group, obtain a result.
Herein, equation only needs to add newline in the matrix bottom also widenable to multiple-camera, adds a plurality of video camera equations, and equation form is identical.
2. if do not make small angle approximation, need two video cameras equally, system of equations is:
x 1 ′ - p 14 y 1 ′ - p 24 . . . . . . . x n ′ - pp 14 y n ′ - pp 24 = X 1 p 11 X 1 p 12 X 1 p 13 Y 1 p 11 Y 1 p 12 Y 1 p 13 Z 1 p 11 Z 1 p 12 Z 1 p 13 p 11 p 12 p 13 X 1 p 21 X 1 p 22 X 1 p 23 Y 1 p 21 Y 1 p 22 Y 1 p 23 Z 1 p 21 Z 1 p 22 Z 1 p 23 p 21 p 22 p 23 . . . . . . . X n pp 11 X n pp 12 X n pp 13 Y n pp 11 Y n pp 12 Y n pp 13 Z n pp 11 Z n pp 12 Z n pp 13 pp 11 pp 12 pp 13 X n pp 21 X n pp 22 X n pp 23 Y n pp 21 Y n pp 22 Y n pp 23 Z n pp 21 Z n pp 22 Z n pp 23 pp 21 pp 22 pp 23 r 11 r 21 r 31 r 12 r 22 r 32 r 13 r 23 r 33 t 1 t 2 t 3 - - - ( 13 )
For non-approximate data, in order to guarantee the matrix A full rank, need satisfy m 〉=4 and n 〉=4, altogether the minimum 4+4=8 group of feature point that needs.
Equally, equation is also widenable to multiple-camera.
2.2 step 2. is found the solution the least mean-square error of system of linear equations b=Ax and is separated
Adopt QR (orthonomal matrix-upper triangular matrix) decomposition method, solution is as follows:
Since the existence of error, Ax=b+ ε, problem is to find the solution x to make norm || ε || 2 2Minimum.
Can find Q to make QA = R O . Wherein Q is an orthogonal matrix, and R is nonsingular upper triangular matrix.
| | ϵ | | 2 2 = | | Ax - b | | 2 2 = | | Q ( Ax - b ) | | 2 2 = | | QAx - Qb | | 2 2 = | | R O - Qb | | 2 2
Note Qb = b 1 b 2 , | | Ax - b | | 2 2 = | | Rx - b b 2 | | 2 2 . Because be column vector, so=|| Rx-b 1|| 2 2+ || b 2|| 2 2
Notice || b 2|| 2 2Be constant, so former problem || Ax-b|| 2 2Min is converted into || Rx-b 1|| 2 2Min, that is: Rx-b 1=0.
So have:
x=R -1b 1 (14)
So far, the x vector solves.
To the approximate data of system of equations (12), x=(φ 1, φ 2, φ 3, t 1, t 2, t 3) T,
To the non-approximate data of system of equations (13), x=(r 11, r 21, r 31, r 12, r 22, r 32, r 13, r 23, r 33, t 1, t 2, t 3) T.
Three, the three-dimensional motion vector of calculating according to step 2 is calculated the target three-dimensional model of key frame, to two video cameras, the two dimensional surface of this video camera matrix correspondence returns in back projection respectively, compare with the key frame angle point two-dimensional coordinate of this video camera, reject the point of error beyond threshold value, two calculate three-dimensional motion vector again set by step then.
The mistake coupling can cause back projection's error very big, and this error directly influences result of calculation, so that the present invention proposes to reject is overdue.Because back projection's error that mistake coupling causes is generally at 10 more than the pixel, even 20, and general normal point is many in 2 pixels, is 10 so establish threshold value.
As long as surpass threshold value in x direction or y deflection error, promptly as overdue rejecting.(the individual point of l≤n) carries out step 2 again, recomputates three-dimensional motion vector by (15) formula or (16) formula for the l that keeps then.
x 1 ′ - p 14 - X 1 p 11 - Y 1 p 12 - Z 1 p 13 y 1 ′ - p 24 - X 1 p 21 - Y 1 p 22 - Z 1 p 23 M M x l ′ - pp 14 - X l pp 11 - Y l pp 12 - Z l pp 13 y l ′ - pp 24 - X l pp 21 - Y l pp 22 - Z l pp 23 = Y 1 p 13 - Z 1 p 12 Z 1 p 11 - X 1 p 13 X 1 p 12 - Y 1 p 11 p 11 p 12 p 13 Y 1 p 23 - Z 1 p 22 Z 1 p 21 - X 1 p 23 X 1 p 22 - Y 1 p 21 p 21 p 22 p 23 M M Y l pp 13 - Z l pp 12 Z l pp 11 - X l pp 13 X l pp 12 - Y l pp 11 pp 11 pp 12 pp 13 Y l pp 23 - Z l pp 22 Z l pp 21 - X l pp 23 X l pp 22 - Y l pp 21 pp 21 pp 22 pp 23 φ 1 φ 2 φ 3 t 1 t 2 t 3 - - - ( 15 )
x 1 ′ - p 14 y 1 ′ - p 24 . . . . . . . x l ′ - pp 14 y l ′ - pp 24 = X 1 p 11 X 1 p 12 X 1 p 13 Y 1 p 11 Y 1 p 12 Y 1 p 13 Z 1 p 11 Z 1 p 12 Z 1 p 13 p 11 p 12 p 13 X 1 p 21 X 1 p 22 X 1 p 23 Y 1 p 21 Y 1 p 22 Y 1 p 23 Z 1 p 21 Z 1 p 22 Z 1 p 23 p 21 p 22 p 23 . . . . . . . X l pp 11 X l pp 12 X l pp 13 Y l pp 11 Y l pp 12 Y l pp 13 Z l pp 11 Z l pp 12 Z l pp 13 pp 11 pp 12 pp 13 X l pp 21 X l pp 22 X l pp 23 Y l pp 21 Y l pp 22 Y l pp 23 Z l pp 21 Z l pp 22 Z l pp 23 pp 21 pp 22 pp 23 r 11 r 21 r 31 r 12 r 22 r 32 r 13 r 23 r 33 t 1 t 2 t 3
(16)
Think that the result is final three-dimensional motion vector.
Comparatively ripe Han-Kanade method in the present invention and the existing research has been done the control experiment.The Han-Kanade method is based on the three-dimensional reconstruction of uncalibrated image not, also can recover motion.Experiment embodiment is: real-world object has rotated 5 degree around the x direction, and the control input point is identical, obtains calculating the anglec of rotation to such as Fig. 3.Precision is in same level as can be seen.And on the time, MATLAB realizes that the computing of one time three frame needs 5~10 minutes, and comprehensive 360 degree are rebuild object just needs the more time.5 video cameras needed more than half an hour.And video camera is few more, and is high more to the requirement of wide-angle coupling, so can not reduce video camera simply.Therefore, in the free-viewing angle display system, can carry out three-dimensional reconstruction consuming time every the long period, and calculate motion in real time, adjust the position of three-dimensional body, make to show that real-time is better, not sacrifice precision simultaneously with the inventive method.
The invention provides the thinking and the method for three-dimensional vectors computing method in a kind of Stereoscopic Video Presentation; the method and the approach of this technical scheme of specific implementation are a lot; the above only is a preferred implementation of the present invention; should be understood that; for those skilled in the art; under the prerequisite that does not break away from the principle of the invention, can also make some improvements and modifications, these improvements and modifications also should be considered as protection scope of the present invention.The all available prior art of each ingredient not clear and definite in the present embodiment is realized.

Claims (4)

1. target three-dimensional motion vector computing method is characterized in that, may further comprise the steps:
Step (1), target Corner Detection and coupling: the target in the identification video stream marks out the angle point of target in basic frame and the key frame, and mates;
Step (2) according to the angle point that mates in basic frame and the key frame, is calculated the three-dimensional motion vector of key frame with respect to basic frame;
Step (3) is calculated the target three-dimensional model of key frame according to basic three-dimensional model and described three-dimensional motion vector, and two dimensional surface returns in back projection with the target three-dimensional model, and rejects overduely, calculates the final goal three-dimensional motion vector.
2. a kind of target three-dimensional motion vector computing method according to claim 1 is characterized in that step (1) may further comprise the steps:
Target in the identification video stream in key frame and the basic frame;
Mark out the angle point of target in basic frame and the key frame;
Angle point to basic frame and key frame mates: find the corresponding angle point of key frame and basic frame, obtain the two-dimensional coordinate of angle point.
3. a kind of target three-dimensional motion vector computing method according to claim 2 is characterized in that step (2) may further comprise the steps:
Based on video camera 1 and video camera 2:
Use approximation method to draw system of linear equations:
x 1 ′ - p 14 - X 1 p 11 - Y 1 p 12 - Z 1 p 13 y 1 ′ - p 24 - X 1 p 21 - Y 1 p 22 - Z 1 p 23 M M x n ′ - pp 14 - X n pp 11 - Y n pp 12 - Z n pp 13 y n ′ - pp 24 - X n pp 21 - Y n pp 22 - Z n pp 23 = Y 1 p 13 - Z 1 p 12 Z 1 p 11 - X 1 p 13 X 1 p 12 - Y 1 p 11 p 11 p 12 p 13 Y 1 p 23 - Z 1 p 22 Z 1 p 21 - X 1 p 23 X 1 p 22 - Y 1 p 21 p 21 p 22 p 23 M M Y n pp 13 - Z n pp 12 Z n pp 11 - X n pp 13 X n pp 12 - Y n pp 11 pp 11 pp 12 pp 13 Y n pp 23 - Z n pp 22 Z n pp 21 - X n pp 23 X n pp 22 - Y n pp 21 pp 21 pp 22 pp 23 φ 1 φ 2 φ 3 t 1 t 2 t 3 ,
φ 1, φ 2, φ 3Be respectively the three-dimensional motion vector reduced parameter, t 1, t 2, t 3Be respectively the translation matrix parameter; Form three-dimensional motion vector M:
M = 1 - φ 3 φ 2 t 1 φ 3 1 - φ 1 t 2 - φ 2 φ 1 1 t 3 0 0 0 1 ;
Or use non-approximation method, draw system of linear equations:
x 1 ′ - p 14 y 1 ′ - p 24 M M M M M M M x n ′ - pp 14 y n ′ - pp 24 = X 1 p 11 X 1 p 12 X 1 p 13 Y 1 p 11 Y 1 p 12 Y 1 p 13 Z 1 p 11 Z 1 p 12 Z 1 p 13 p 11 p 12 p 13 X 1 p 21 X 1 p 22 X 1 p 23 Y 1 p 21 Y 1 p 22 Y 1 p 23 Z 1 p 21 Z 1 p 22 Z 1 p 23 p 21 p 22 p 23 . M M M M M M X n pp 11 X n pp 12 X n pp 13 Y n pp 11 Y n pp 12 Y n pp 13 Z n pp 11 Z n pp 12 Z n pp 13 pp 11 pp 12 pp 13 X n pp 21 X n pp 22 X n pp 23 Y n pp 21 Y n pp 22 Y n pp 23 Z n pp 21 Z n pp 22 Z n pp 23 pp 21 pp 22 pp 23 r 11 r 21 r 31 r 12 r 22 r 32 r 13 r 23 r 33 t 1 t 2 t 3
, r wherein 11, r 21, r 31, r 12, r 22, r 32, r 13, r 23, r 33Be the rotation matrix parameter, form three-dimensional motion vector M:
M = r 11 r 12 r 13 t 1 r 21 r 22 r 23 t 2 r 31 r 32 r 33 t 3 0 0 0 1 ;
(X wherein i, Y i, Z i) any some three-dimensional coordinate in the basic frame of video camera 1 or video camera 2 of expression; (X i', Y i', Z i') represent this three-dimensional coordinate in the key frame of video camera 1 or video camera 2; (x i, y i) represent that this projects to the two-dimensional coordinate of two dimensional surface at video camera 1 or video camera 2 in basic frame; (x i', y i') represent that this projects to the two-dimensional coordinate of two dimensional surface at video camera 1 or video camera 2 in key frame;
P = p 11 p 12 p 13 p 14 p 21 p 22 p 23 p 24 0 0 0 1 Projection matrix for video camera 1;
PP = pp 11 pp 12 pp 13 pp 14 pp 21 pp 22 pp 23 pp 24 0 0 0 1 Projection matrix for video camera 2;
Find the solution the least mean-square error of system of linear equations and separate, obtain the kinematic matrix parameter, form three-dimensional motion vector.
4. a kind of target three-dimensional motion vector computing method according to claim 3 is characterized in that step (3) may further comprise the steps:
To two video cameras, described kinematic matrix parameter is additional to basic three-dimensional model, obtain the three-dimensional model of two camera video key frames;
Respectively with the three-dimensional model of two camera video key frames, the two dimensional surface of video camera projection matrix correspondence separately returns in back projection, with the angle point two-dimensional coordinate of the key frame of corresponding video camera relatively, reject the point of error beyond threshold value, use step (2) to recomputate the three-dimensional motion matrix parameter, obtain final three-dimensional motion vector.
CN 201110176736 2011-06-28 2011-06-28 Method for calculating three-dimensional object motion vector Expired - Fee Related CN102222348B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN 201110176736 CN102222348B (en) 2011-06-28 2011-06-28 Method for calculating three-dimensional object motion vector

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN 201110176736 CN102222348B (en) 2011-06-28 2011-06-28 Method for calculating three-dimensional object motion vector

Publications (2)

Publication Number Publication Date
CN102222348A true CN102222348A (en) 2011-10-19
CN102222348B CN102222348B (en) 2013-04-24

Family

ID=44778892

Family Applications (1)

Application Number Title Priority Date Filing Date
CN 201110176736 Expired - Fee Related CN102222348B (en) 2011-06-28 2011-06-28 Method for calculating three-dimensional object motion vector

Country Status (1)

Country Link
CN (1) CN102222348B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104123716A (en) * 2013-04-28 2014-10-29 腾讯科技(深圳)有限公司 Image stability detection method, device and terminal
US9317770B2 (en) 2013-04-28 2016-04-19 Tencent Technology (Shenzhen) Co., Ltd. Method, apparatus and terminal for detecting image stability
CN107230220A (en) * 2017-05-26 2017-10-03 深圳大学 A kind of new space-time Harris angular-point detection methods and device
CN107707899A (en) * 2017-10-19 2018-02-16 中科创达软件股份有限公司 Multi-view image processing method, device and electronic equipment comprising moving target
CN107808388A (en) * 2017-10-19 2018-03-16 中科创达软件股份有限公司 Image processing method, device and electronic equipment comprising moving target
CN109063567A (en) * 2018-07-03 2018-12-21 百度在线网络技术(北京)有限公司 Human body recognition method, device and storage medium
CN109146932A (en) * 2018-07-17 2019-01-04 北京旷视科技有限公司 Determine the methods, devices and systems of the world coordinates of target point in image
CN111179271A (en) * 2019-11-22 2020-05-19 浙江众合科技股份有限公司 Object angle information labeling method based on retrieval matching and electronic equipment
US11920968B2 (en) 2018-03-06 2024-03-05 Ebm-Papst Mulfingen Gmbh & Co. Kg Device and method for air quantity recording

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040165781A1 (en) * 2003-02-19 2004-08-26 Eastman Kodak Company Method and system for constraint-consistent motion estimation
CN101336856A (en) * 2008-08-08 2009-01-07 西安电子科技大学 Information acquisition and transfer method of auxiliary vision system
CN101729920A (en) * 2009-11-23 2010-06-09 南京大学 Method for displaying stereoscopic video with free visual angles

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040165781A1 (en) * 2003-02-19 2004-08-26 Eastman Kodak Company Method and system for constraint-consistent motion estimation
CN101336856A (en) * 2008-08-08 2009-01-07 西安电子科技大学 Information acquisition and transfer method of auxiliary vision system
CN101729920A (en) * 2009-11-23 2010-06-09 南京大学 Method for displaying stereoscopic video with free visual angles

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104123716A (en) * 2013-04-28 2014-10-29 腾讯科技(深圳)有限公司 Image stability detection method, device and terminal
US9317770B2 (en) 2013-04-28 2016-04-19 Tencent Technology (Shenzhen) Co., Ltd. Method, apparatus and terminal for detecting image stability
CN107230220A (en) * 2017-05-26 2017-10-03 深圳大学 A kind of new space-time Harris angular-point detection methods and device
CN107707899A (en) * 2017-10-19 2018-02-16 中科创达软件股份有限公司 Multi-view image processing method, device and electronic equipment comprising moving target
CN107808388A (en) * 2017-10-19 2018-03-16 中科创达软件股份有限公司 Image processing method, device and electronic equipment comprising moving target
CN107808388B (en) * 2017-10-19 2021-10-12 中科创达软件股份有限公司 Image processing method and device containing moving object and electronic equipment
US11920968B2 (en) 2018-03-06 2024-03-05 Ebm-Papst Mulfingen Gmbh & Co. Kg Device and method for air quantity recording
KR20200068709A (en) * 2018-07-03 2020-06-15 바이두 온라인 네트웍 테크놀러지 (베이징) 캄파니 리미티드 Human body identification methods, devices and storage media
WO2020007156A1 (en) * 2018-07-03 2020-01-09 百度在线网络技术(北京)有限公司 Human body recognition method and device, as well as storage medium
CN109063567B (en) * 2018-07-03 2021-04-13 百度在线网络技术(北京)有限公司 Human body recognition method, human body recognition device and storage medium
EP3819815A4 (en) * 2018-07-03 2022-05-04 Baidu Online Network Technology (Beijing) Co., Ltd. Human body recognition method and device, as well as storage medium
US11354923B2 (en) 2018-07-03 2022-06-07 Baidu Online Network Technology (Beijing) Co., Ltd. Human body recognition method and apparatus, and storage medium
CN109063567A (en) * 2018-07-03 2018-12-21 百度在线网络技术(北京)有限公司 Human body recognition method, device and storage medium
CN109146932A (en) * 2018-07-17 2019-01-04 北京旷视科技有限公司 Determine the methods, devices and systems of the world coordinates of target point in image
CN111179271A (en) * 2019-11-22 2020-05-19 浙江众合科技股份有限公司 Object angle information labeling method based on retrieval matching and electronic equipment

Also Published As

Publication number Publication date
CN102222348B (en) 2013-04-24

Similar Documents

Publication Publication Date Title
CN102222348B (en) Method for calculating three-dimensional object motion vector
EP3067861B1 (en) Determination of a coordinate conversion parameter
CN109166077B (en) Image alignment method and device, readable storage medium and computer equipment
CN104933718B (en) A kind of physical coordinates localization method based on binocular vision
CN108399643A (en) A kind of outer ginseng calibration system between laser radar and camera and method
CN103115613B (en) Three-dimensional space positioning method
KR20160121798A (en) Hmd calibration with direct geometric modeling
CN102842117B (en) Method for correcting kinematic errors in microscopic vision system
CN104376552A (en) Virtual-real registering algorithm of 3D model and two-dimensional image
CN106803270A (en) Unmanned aerial vehicle platform is based on many key frames collaboration ground target localization method of monocular SLAM
CN101729920B (en) Method for displaying stereoscopic video with free visual angles
CN110009690A (en) Binocular stereo vision image measuring method based on polar curve correction
WO2016030305A1 (en) Method and device for registering an image to a model
CN110070598A (en) Mobile terminal and its progress 3D scan rebuilding method for 3D scan rebuilding
CN107862733B (en) Large-scale scene real-time three-dimensional reconstruction method and system based on sight updating algorithm
CN111415375B (en) SLAM method based on multi-fisheye camera and double-pinhole projection model
CN112712566A (en) Binocular stereo vision sensor measuring method based on structure parameter online correction
Gaschler et al. Epipolar-based stereo tracking without explicit 3d reconstruction
Kurz et al. Bundle adjustment for stereoscopic 3d
Huang et al. MC-VEO: A Visual-Event Odometry With Accurate 6-DoF Motion Compensation
Golban et al. A Comparison Study on Replacing Stereo Disparity with LiDAR in Visual Odometry Methods
KR102375135B1 (en) Apparatus and Method for Cailbrating Carmeras Loaction of Muti View Using Spherical Object
CN109089100B (en) Method for synthesizing binocular stereo video
Wang et al. Motion-based temporal alignment of independently moving cameras
Zhang et al. A practical calibration method for multiple cameras

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20130424

Termination date: 20160628

CF01 Termination of patent right due to non-payment of annual fee