throbber
Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 1 of 29 PageID #: 80
`
`
`
`Exhibit 2
`
`
`
`
`
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 2 of 29 PageID #: 81
`I 1111111111111111 11111 1111111111 111111111111111 IIIII IIIII IIIIII IIII IIII IIII
`US007336814B2
`
`c12) United States Patent
`Boca et al.
`
`(IO) Patent No.:
`(45) Date of Patent:
`
`US 7,336,814 B2
`Feb.26,2008
`
`(54) METHOD AND APPARATUS FOR
`MACHINE-VISION
`
`(75)
`
`Inventors: Remus F Boca, North Vancouver (CA);
`Babak Habibi, North Vancouver (CA);
`Mohammad Sameti, Coquitlam (CA);
`Simona Pescaru, Vancouver (CA)
`
`4,334,241 A
`4,437,114 A
`4,578,561 A
`4,613,942 A
`
`6/1982 Kashioka et al ............ 358/107
`3/1984 LaRussa ..................... 358/101
`3/1986 Corby, Jr. et al.
`...... 219/124.34
`9/1986 Chen .......................... 364/513
`
`(Continued)
`
`FOREIGN PATENT DOCUMENTS
`
`(73) Assignee: Braintech Canada, Inc., North
`Vancouver (CA)
`
`EP
`
`0 114 505
`
`8/1984
`
`( *) Notice:
`
`Subject to any disclaimer, the term of this
`patent is extended or adjusted under 35
`U.S.C. 154(b) by O days.
`
`(Continued)
`
`OTHER PUBLICATIONS
`
`(21) Appl. No.: 11/183,228
`
`(22) Filed:
`
`Jul. 14, 2005
`
`(65)
`
`Prior Publication Data
`
`US 2006/0088203 Al
`
`Apr. 27, 2006
`
`Related U.S. Application Data
`
`(60) Provisional application No. 60/587,488, filed on Jul.
`14, 2004.
`
`(51)
`
`Int. Cl.
`G06K 9/00
`(2006.01)
`(52) U.S. Cl. ...................... 382/141; 382/142; 382/143;
`382/144; 382/147; 382/152
`(58) Field of Classification Search ........ 382/141-153,
`382/155-160
`See application file for complete search history.
`
`(56)
`
`References Cited
`
`U.S. PATENT DOCUMENTS
`
`3,986,007 A
`4,146,924 A
`4,219,847 A
`4,305,130 A
`
`10/1976
`3/1979
`8/1980
`12/1981
`
`Ruoff, Jr .................. 235/151.1
`Birk et al.
`.................. 364/513
`Pinkney et al.
`............. 358/126
`Kelley et al.
`............... 364/513
`
`"3D Robot Guidance with a Single Camera," ISRA Vision Systems
`AG, pp. 83-105, no date available.
`
`(Continued)
`
`Primary Examiner-Brian Le
`(74) Attorney, Agent, or Firm-Seed IP Law Group PLLC
`
`(57)
`
`ABSTRACT
`
`A system and method facilitate machine-vision, for example
`three-dimensional pose estimation for target objects, using
`one or more images sensors to acquire images of the target
`object at one or more positions, and to identify features of
`the target object in the resulting images. A set of equations
`is set up exploiting invariant physical relationships between
`features such as constancy of distances, angles, and areas or
`volumes enclosed by or between features. The set of equa(cid:173)
`tions may be solved to estimate a 3D pose. The number of
`positions may be determined based on the number of image
`sensors, number of features identified, and/or number of
`known physical relationships between less than all features.
`Knowledge of physical relationships between image sensors
`and/or between features and image sensors may be
`employed. A robot path may be transformed based on the
`pose, to align the path with the target object.
`
`39 Claims, 12 Drawing Sheets
`
`f/24
`
`140
`
`ALL IMAGES
`FROM ALL VIEWS
`
`TAKE NEXT VIEW (from I to m)
`
`TAKE NEXT IMAGE OF CURRENT VIEW
`(from I lo c)
`
`LOCATE FEATURES PREVIOUSLY TRAINED
`CORRESPONDING TO THE IMAGE SENSOR
`THAT ACQUIRED THE
`IMAGE
`
`EXTRACT FEATURE INFORMATION
`
`NO
`GO TO COMPUTE POSITION
`
`150
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 3 of 29 PageID #: 82
`
`US 7,336,814 B2
`Page 2
`
`U.S. PATENT DOCUMENTS
`4/1987 Pryor .......................... 29/407
`4,654,949 A
`5/1989 Suzuki .................. 318/568.13
`4,835,450 A
`7/1990 McGee et al. .............. 364/513
`4,942,539 A
`5,083,073 A
`1/1992 Kato .......................... 318/577
`5,212,738 A *
`5/1993 Chande et al. .............. 382/152
`9/1994 Azuma et al.
`.............. 414/416
`5,350,269 A
`5,454,775 A
`10/1995 Cullen et al. ................. 483/16
`12/1997 Pryor .................... 364/167.01
`5,696,673 A
`9/1999 Pryor ......................... 382/154
`5,956,417 A
`9/1999 Bieman et al. ........ 318/568.15
`5,959,425 A
`5,978,521 A * 11/1999 Wallack et al. ............. 382/294
`6,044,183 A
`3/2000 Pryor ......................... 382/287
`6,115,480 A *
`9/2000 Washizawa ................. 382/103
`6,141,863 A
`11/2000 Hara et al. .................... 29/714
`1/2001 Pryor ...................... 29/407.04
`6,167,607 Bl
`6,211,506 Bl
`4/2001 Pryor et al. .............. 250/208.1
`6,301,763 Bl
`10/2001 Pryor ...................... 29/407.04
`6,341,246 Bl
`1/2002 Gerstenberger et al.
`.... 700/245
`6,466,843 Bl
`10/2002 Bonanni et al.
`............ 700/245
`6,594,600 Bl
`7/2003 Arnoul et al.
`................ 702/94
`9/2003 Huang et al. ............... 382/154
`6,628,819 Bl*
`6/2004 Fujita et al. ................ 700/245
`6,754,560 B2
`10/2004 Bachelder et al. .......... 382/294
`6,804,416 Bl
`11/2004 Habibi et al. ............... 700/259
`6,816,755 B2
`6/2006 Franke et al. ............... 356/604
`7,061,628 B2 *
`1/2003 Aliaga et al. .................. 703/2
`2003/0004694 Al
`2003/0007159 Al *
`1/2003 Franke et al. ............... 356/604
`2004/0114033 Al
`6/2004 Eian et al. .................... 348/42
`2004/0172164 Al
`9/2004 Habibi et al.
`
`FOREIGN PATENT DOCUMENTS
`
`EP
`EP
`EP
`EP
`EP
`JP
`
`0151417
`0493612
`0763406 Bl
`0911603 Bl
`0951968
`1-124072
`
`8/1985
`7 /1992
`9/1997
`4/1999
`10/1999
`5/1989
`
`OTHER PUBLICATIONS
`
`3D Vision with One Camera, URL~http://neu.isravision.com/
`likecms/index.php?site~site.html&diFisra&nav~ 162,
`download
`date Apr. 12, 2005.
`Bejczy, A. K., "Challenges of Human-Robot Communication in
`Telerobotics," IEEE International Workshop on Robot and Human
`Communication, pp. 1-8, Nov. 11-14, 1996.
`Borotschnig, H., et al., "Appearance-Based Active Object Recog(cid:173)
`nition," Image and Vision Computing, 18:715-727, 2000.
`U.S. Appl. No. 60/413,180, filed Sep. 23, 2002, Eian et al.
`
`U.S. Appl. No. 60/587,488, filed Jul. 14, 2004, Boca et al.
`Denzler, J., et al., "Learning, Tracking and Recognition of 3D
`Objects," Proceedings of the International Conference on Intelli(cid:173)
`gent Robots and Systems (!ROS), XP000512662, 1:89-96, Munich,
`Sep. 12, 1994 .
`Huang, T., et al., "Uniqueness of3D Pose Under Weak Perspective:
`A Geometrical Proof," IEEE Transactions on Pattern Analysis and
`Machine Intelligence, 17(12):1220-1221, Dec. 1995.
`Ji, Q., et al., "An Integrated Linear Technique for Pose Estimation
`from Different Geometric Features," International Journal of Pattern
`Recognition and Artificial Intelligence, 13(5):705-733, Aug. 1999.
`Jia, Y-B., et al., "Sensing Polygon Poses by Inscription," in Pro(cid:173)
`ceedings of 1994 IEEE International Conference on Robotics and
`Automation, Los Alamitos, CA, May 8, 1994, pp. 1642-1649.
`Kim, W., "Computer Vision Assisted Virtual Reality Calibration,"
`URL~http://www-robotics.jpl.nasa.gov/publications/Won_Kirn/
`ra98_vrc.pdf.
`Kovacic, S., et al., "Planning Sequences of Views for 3-D Object
`Recognition and Pose Determination," Pattern Recognition,
`31(10):1407-1417, 1998.
`Liu, Y., et al., "Determination of Camera Location from 2D to 3D
`Line and Point Correspondences", IEEE Transaction on Pattern
`Analysis and Machine Intelligence, 12(1), Jan. 1990.
`Lu, C-P., et al., "Fast and Globally Convergent Pose Estimation
`from Video Images," Transactions on Pattern Analysis and
`Machine Intelligence, 22(6):610-622, Jun. 2000.
`Meyer, W.,"One-Eyed Robots With 3D Vision," ISRA Vision Sys(cid:173)
`tems AG, Press News, Release No. 16, Jun. 15, 2004, pp. 1-7 .
`Sanchez, A., et al., "Robot-Arm Pick and Place Behavior Program(cid:173)
`ming System Using Visual Perception," in Proceedings of the 15th
`International Conference on Pattern Recognition, Los Alamitos,
`CA, Sep. 3-7, 2000, 4:507-510.
`Sharma, R., "Visual Servoing with Independently Controlled Cam(cid:173)
`eras Using a Learned Invariant Representation," in Proceedings of
`the 37"' IEEE Conference on Decision & Control, Tampa, FL, Dec.
`1998, pp. 3263-3268.
`Tsai, R., et al., "A New Technique for Fully Autonomous and
`Efficient 3D Robotics Hand/Eye Calibration," in IEEE Transactions
`on Robotics and Automation, 5(3):345-358, Jun. 1989.
`Wei, G.-Q., et al., "Active Self-Calibration of Robotic Eyes and
`Hand-Eye Relationships With Model Identification," IEEE Trans(cid:173)
`actions on Robotics and Automation, 14(1):158-166, Feb. 1998.
`Wei, G-Q., et al., "Multisensory Visual Servoing By a Neural
`Network," IEEE Transactions on Systems, Man and Cybernetics,
`Part B: CYBERNETICS, 29(2):276-280, Apr. 1999.
`Zhang, Z., "A Flexible New Technique for Camera Calibration,"
`URL~http://research.microsoft.com/research/pubs/view.
`aspx?tr_id~212.
`
`* cited by examiner
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 4 of 29 PageID #: 83
`
`U.S. Patent
`
`Feb.26,2008
`
`Sheet 1 of 12
`
`US 7,336,814 B2
`
`~
`
`~
`
`' ' '
`
`'
`
`~
`~
`I
`\, ,,,. .. /'
`,,"'
`
`/ /
`
`,,"
`
`,~
`
`~ ~
`
`~ co ;r
`
`~
`~
`
`~
`
`~
`"'or::t-
`
`C:) -)
`
`' ' ', _,._
`..-
`
`/
`I
`I
`~
`~
`
`'
`
`'
`
`' ,
`,
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 5 of 29 PageID #: 84
`
`"'""' ~ = N
`00
`0--,
`-....l w w
`d r.,;_
`
`....
`0 ....
`N
`.....
`rJJ =- ('D
`
`('D
`
`N
`
`QO
`0
`0
`N
`~Cl's
`N
`?'
`('D
`"f'j
`
`~ = ~
`
`~
`~
`~
`•
`00
`
`e •
`
`'--10
`
`I
`
`40a
`
`I
`
`I
`
`20
`
`18
`\
`
`li
`
`FIG. 2A
`
`I
`
`I
`
`I
`
`I
`
`14
`
`42b
`
`\__
`
`42c
`
`I
`
`\
`
`24a
`
`42g/9
`
`42i
`
`~42f
`
`42d
`
`/
`
`120---.....
`
`~~42e
`
`42h
`
`42c
`
`24c
`
`40b
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 6 of 29 PageID #: 85
`
`"'""' ~ = N
`00
`0--,
`-....l w w
`d r.,;_
`
`N
`
`~
`
`('D
`('D
`
`....
`0 ....
`.....
`rJJ =(cid:173)
`
`QO
`0
`0
`N
`~Cl's
`N
`?'
`('D
`"f'j
`
`~ = ~
`
`~
`~
`~
`•
`00
`
`e •
`
`'--10
`
`40a
`
`20
`
`FIG. 2B
`
`14
`
`24a
`
`44
`
`12a
`
`42h
`
`42f
`
`42d
`
`;r;::--42e
`
`24c
`
`-40b
`
`24b
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 7 of 29 PageID #: 86
`
`U.S. Patent
`
`Feb.26,2008
`
`Sheet 4 of 12
`
`US 7,336,814 B2
`
`..Q
`~
`
`;§
`
`<..>
`
`~
`~ ~
`
`t:,
`~
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 8 of 29 PageID #: 87
`
`U.S. Patent
`
`Feb.26,2008
`
`Sheet 5 of 12
`
`US 7,336,814 B2
`
`CALIBRATE IMAGE SENSOR(S)
`
`1100
`
`102
`
`----------
`
`TRAIN SYSTEM
`
`r---.- 104
`
`POSE ESTIMATION
`(DETERMINE TARGET OBJECT'S 3-D POSE) ~ 106
`
`FIG. 4
`
`1102
`
`INTRINSIC PARAMETER FOR
`DETERMINE
`EACH
`IMAGE SENSOR
`
`110
`
`----------
`
`DETERMINE POSE OF ONE OR MORE
`IMAGE SENSORS
`
`112
`
`----------
`
`FIG. 5
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 9 of 29 PageID #: 88
`
`U.S. Patent
`
`Feb.26,2008
`
`Sheet 6 of 12
`
`US 7,336,814 B2
`
`1104
`
`ACQUIRE ONE VIEW AND TRAIN FEATURES ~120
`
`DETERMINE NUMBER OF ADDITIONAL VIEWS
`
`---..__,,121
`
`CHANGE RELATIVE OBJECT POSE TO
`IMAGE SENSOR(S) (RIGID) AND ACQUIRE ~122
`VIEW(S)
`
`EXTRACT FEATURES
`
`~124
`
`/128
`
`COMPUTE 3D FEATURE POSITIONS IN
`RESPECTIVE
`IMAGE SENSOR COORDINATE
`FRAMES (CREATE LOCAL MODELS)
`
`126
`
`/
`SPARSE MODEL {
`INFORMATION
`
`I
`I
`
`FIG. 6
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 10 of 29 PageID #: 89
`
`"'""' ~ = N
`00
`0--,
`-....l w w
`d r.,;_
`
`....
`0 ....
`-....J
`.....
`rJJ =- ('D
`
`('D
`
`N
`
`QO
`0
`0
`N
`~Cl's
`N
`?'
`('D
`"f'j
`
`~ = ~
`
`~
`~
`~
`•
`00
`
`e •
`
`FIG. 7
`
`I
`
`GO TO COMPUTE POSITION
`
`j= 1 .. c
`k=1 .. m
`i=1 .. nf
`Fi vk,cj (x,y)
`STORE
`
`150
`
`>
`IMAGES FOR CURRENT VIEW? >
`
`,,.-f 56
`
`cts4
`
`152
`
`l
`
`NO
`
`MORE VIEWS?
`
`t NO
`
`I
`
`YES (
`
`MORE
`
`YES (
`
`148
`
`146
`
`144
`
`INFORMATION
`
`EXTRACT FEATURE
`
`CORRESPONDING TO THE
`IMAGE SENSOR
`LOCATE FEATURES PREVIOUSLY TRAINED
`
`IMAGE
`
`THAT ACQUIRED THE
`
`(from 1 to c)
`IMAGE OF CURRENT VIEW
`
`TAKE NEXT
`
`l'---142
`
`TAKE NEXT VIEW (from 1 to m)
`
`FROM ALL VIEWS
`
`ALL IMAGES
`
`1124
`
`140
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 11 of 29 PageID #: 90
`
`~
`a
`
`O
`QO
`
`g2
`
`QO
`0
`0
`N
`~a-.
`~
`O"
`('D
`"'f"l
`
`""""
`
`=
`;-
`~
`~
`e
`
`"'""' ~ = N
`00
`0--,
`-....l w w
`d r.,;_
`
`FIG. ;A ---J---------------------"
`
`-----------
`
`Fi vm,cc ( X, y) IMAGE COORDINATES .••
`
`LIST OF FEATURES INFORMATION FROM
`
`VIEW m, CAMERA c:
`
`•
`
`~ L
`,,-1
`I
`I
`I
`I
`I
`
`•
`
`•
`
`16~c
`/
`
`A
`
`i
`
`:
`______ _j
`
`____ J________________
`
`----
`
`L_ --------
`
`Fi vm,c1 ( X, y) IMAGE COORDINATES ...
`
`i
`
`I
`I
`
`i
`
`1
`
`FEATURES EXTRAcf rn FROM CAMERA c
`
`IMAGE COORDINATES...
`
`Fiv2,cc(x,y)
`
`LIST OF FEATURES INFORMATION FROM
`
`VIEW 2, CAMERA c:
`
`VIEW m
`
`i
`
`I
`I
`
`i
`• : • • • • •
`
`: VIEW 2
`
`I
`I
`
`LIST OF FEATURES INFORMATION FROM
`
`VIEW m, CAMERA 1:
`
`FEATURES EXTRAcfrn FROM CAMERA 1
`
`IMAGE COORDINATES...
`
`Fiv2,c1(x,y)
`
`/
`
`INFORMATION FROM
`
`LIST OF FEATURES
`
`VIEW 2, CAMERA 1:
`
`I
`'---------------' I
`1
`
`:
`:
`
`I
`I
`I
`I
`
`:
`:
`
`I
`I
`
`I
`I
`I
`I
`
`:
`:
`
`:
`I
`I
`I
`
`VIEW 1 CAMERA 1·
`
`'
`
`1
`
`FEATURES FROM
`
`f I v1 ,c1 (x y)
`
`•
`
`'
`
`I
`----------'--~ I
`
`IMAGE COORDINATES...
`
`fiv1,cc(x,y)
`
`:
`
`:
`
`IMAGE COORDINATES...
`
`Fiv1,c1(x,y)
`
`1
`I ~------------~ I
`
`LIST OF FEATURES INFORMATION FROM
`
`VIEW 1, CAMERA c:
`
`• : • • • • : •
`: VIEW 1 :
`
`~..,
`
`INFORMATION FROM
`
`LIST OF FEATURES
`
`VIEW 1, CAMERA 1:
`
`I
`I
`,--------------------,
`
`CAMERA c
`
`l /;
`1600
`
`CAMERA 1
`
`,,,--128
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 12 of 29 PageID #: 91
`
`....
`0 ....
`.....
`rJJ =(cid:173)
`
`1,0
`
`('D
`('D
`
`QO
`0
`0
`N
`~Cl's
`N
`?'
`('D
`"f'j
`
`~ = ~
`
`~
`~
`~
`•
`00
`
`e •
`
`"'""' ~ = N
`00
`0--,
`-....l w w
`d r.,;_
`
`FIG. BB
`
`N
`
`n=1 .. TOTAL NUMBER OF FEATURES FOR ALL IMAGE SENSOR(S)
`b=1 .. c
`a=1 .. m
`COORDINATE SYSTEMS
`RELATIVE TO THE RESPECTIVE
`IMAGE SENSOR
`P: 30 POSITION (x,y,z) OF A GIVEN FEATURE
`WHERE:
`pn va, cb
`
`170
`
`P(FEA TURE j, FEATURE i)
`
`SPARSE MODEL INFORMATION:
`
`126
`
`168
`
`FEATURES AND/OR IMAGE SENSOR(S)
`
`INFORMATION OF ALL THE
`
`THE 30 POSITION
`
`MARQUARDT) TO SOLVE THE ABOVE EQUATIONS FOR
`USE A MINIMIZATION ALGORITHM (LIKE LEVENBNERG
`
`164
`
`P(Fi vs,cp, ... Fi vs,cr) = P(Fi vt,cp, ... Fi vt,cr)
`
`IN ANOTHER VIEW
`
`PHYSICAL PARAMETERS
`
`ANO OTHER FEATURES
`IN ONE VIEW WITH THE SAME
`THE PHYSICAL PARAMETERS BETWEEN EACH FEATURE
`CONSTRUCT NONLINEAR EQUATIONS SET BY EQUATING
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 13 of 29 PageID #: 92
`
`"'""' ~ = N
`00
`0--,
`-....l w w
`d r.,;_
`
`....
`0 ....
`0
`....
`.....
`rJJ =(cid:173)
`
`('D
`('D
`
`N
`
`QO
`0
`0
`N
`~Cl's
`N
`?'
`('D
`"f'j
`
`~ = ~
`
`~
`~
`~
`•
`00
`
`e •
`
`FIG. 9
`
`204
`
`202
`
`DETERMINE 3D OBJECT POSE
`
`USE TRANSFORMED FEATURES POSITIONS TO
`
`FRAME TO THE COMMON REFERENCE FRAME
`
`DESIRED VIEW ( p ... va, c ... ) FROM EACH CAMERA
`
`TRANSFORM FEATURES POSITIONS FOR THE
`
`n=1 .. TOTAL NUMBER OF FEATURES FOR ALL IMAGE SENSOR(S)
`b=1 .. c
`a=1 .. m
`COORDINATE SYSTEMS
`RELATIVE TO THE RESPECTIVE
`IMAGE SENSOR
`P: 3D POSITION (x,y,z) OF A GIVEN FEATURE
`WHERE:
`pn va, cb
`
`106a
`
`f
`
`200
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 14 of 29 PageID #: 93
`
`U.S. Patent
`
`Feb.26,2008
`
`Sheet 11 of 12
`
`US 7,336,814 B2
`
`220
`
`1 106b
`
`pnva, cb
`WHERE:
`P: 3D POSITION (x,y,z) OF A GIVEN FEATURE
`RELATIVE TO THE RESPECTIVE
`IMAGE SENSOR
`COORDINATE SYSTEMS
`a=1 .. m
`b=1 .. c
`n=1 .. TOTAL NUMBER OF FEATURES FOR ALL IMAGE
`SENSOR(S)
`
`SELECT AT LEAST THREE NON COLINEAR
`FEATURES Fi 1 Fi 2 Fi3
`
`I
`
`I
`
`224a
`
`222
`
`224b
`
`TAKE SELECTED FEATURES
`Fi2
`Fi3
`( Fil
`va,c .. ,
`va,c .. ,
`va,c ..
`FROM THE VIEW THE OBJECT WAS
`IN THE OBJECT REFERENCE POSITION
`
`)
`
`TAKE SELECTED FEATURES
`Fi2
`Fi3
`( Fi1
`vb,c .. ,
`vb,c .. ,
`vb,c ..
`FROM THE VIEW THE OBJECT
`WAS
`IN THE CURRENT POSITION
`
`)
`
`226a
`
`226b
`
`TRANSFORM SELECTED FEATURES
`POSITIONS (P·" VO, c ... )
`FROM EACH CAMERA FRAME
`TO THE COMMON REFERENCE FRAME
`
`TRANSFORM SELECTED FEATURES
`POSITIONS ( p .. · vb, c ... )
`FROM EACH CAMERA FRAME
`TO THE COMMON REFERENCE FRAME
`
`228a
`
`228b
`
`IN THE SAME
`USE THESE POINTS
`COMMON REFERENCE FRAME TO
`BUILD OBJECT REFERENCE
`FRAME
`
`IN THE SAME
`USE THESE POINTS
`COMMON REFERENCE FRAME TO
`BUILD CURRENT OBJECT FRAME
`
`FIG. 10
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 15 of 29 PageID #: 94
`
`U.S. Patent
`
`Feb.26,2008
`
`Sheet 12 of 12
`
`US 7,336,814 B2
`
`220
`
`f
`
`106c
`
`pn va, cb
`WHERE:
`P: 3D POSITION (x,y,z) OF A GIVEN FEATURE
`RELATIVE TO THE RESPECTIVE
`IMAGE SENSOR
`COORDINATE SYSTEMS
`a=1 .. m
`b=1 .. c
`n=1 .. TOTAL NUMBER OF FEATURES FOR ALL IMAGE
`SENSOR(S)
`
`SELECT AT LEAST THREE NON COUNEAR
`FEATURES Fi1 Fi2 Fi3
`'
`'
`
`224c
`
`222
`
`224b
`
`TAKE SELECTED FEATURES
`)
`Fi2
`Fi3
`(F i1
`va,c .. ,
`va,c .. ,
`va,c ..
`FROM THE VIEW THE OBJECT WAS
`IN THE PREVIOUS POSITION
`
`TAKE SELECTED FEATURES
`Fi2
`Fi3
`(Fi1
`vb,c .. ,
`vb,c .. ,
`vb,c ..
`FROM THE VIEW THE OBJECT
`WAS
`IN THE CURRENT POSITION
`
`)
`
`226a
`
`226b
`
`TRANSFORM SELECTED FEATURES
`POSITIONS (p ... va, c ... )
`FROM EACH CAMERA FRAME
`TO THE COMMON REFERENCE FRAME
`
`TRANSFORM SELECTED FEATURES
`POSITIONS (p ... vb, c ... )
`FROM EACH CAMERA FRAME
`TO THE COMMON REFERENCE FRAME
`
`228a
`
`228b
`
`IN THE SAME
`USE THESE POINTS
`COMMON REFERENCE FRAME TO
`BUILD OBJECT REFERENCE
`FRAME
`
`IN THE SAME
`USE THESE POINTS
`COMMON REFERENCE FRAME TO
`BUILD CURRENT OBJECT FRAME
`
`FIG. 11
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 16 of 29 PageID #: 95
`
`US 7,336,814 B2
`
`1
`METHOD AND APPARATUS FOR
`MACHINE-VISION
`
`CROSS-REFERENCE TO RELATED
`APPLICATIONS
`
`This application claims benefit under 35 U.S.C. 119(e) to
`U.S. provisional application Ser. No. 60/587,488, filed Jul.
`14, 2004.
`
`BACKGROUND OF THE INVENTION
`
`5
`
`2
`including cameras and lasers, susceptibility to damage or
`misalignment when operating in industrial environments as
`well as posing of a potential safety hazard when laser light
`sources are deployed in proximity of human operators.
`Based on the above considerations it is desirable to devise
`a three-dimensional robot guidance system that eliminates
`the need for stereo camera pairs and the need for the use of
`structured light and specialized sensors. Such a system
`would increase accuracy, simplify setup and maintenance
`10 and reduce hardware costs.
`Prior methods have been developed that utilize a single
`camera to view each region/feature of the object in order to
`calculate the 3D pose of the object. For example see U.S.
`Pat. No. 4,942,539 McGee, and European Patent No.
`15 0911603Bl Ersu. However these and similar methods
`require the calibration of all cameras relative to a common
`coordinate frame such as a robot. In practice such a require(cid:173)
`ment is cumbersome and time-consuming to fulfill and
`difficult to automate. These methods also require a priori
`20 knowledge of the geometrical relationships between all
`object features used. One source for such data is object
`Computer Aided Design (CAD) models; however, such data
`files are often not readily available. In the absence of CAD
`data, past systems have relied on direct object measurement
`25 using a coordinate measurement machine or a robot
`equipped with a pointing device. This process is difficult and
`error prone especially in the case of large objects with
`features that are scattered in different regions.
`It is therefore highly desirable to develop a three-dimen-
`30 sional robot guidance system that in addition to eliminating
`the need for stereo cameras and lasers, also eliminates the
`need for inter-camera calibration and the need for a priori
`knowledge of geometrical relationships between all object
`features.
`
`1. Field of the Invention
`This disclosure relates to the field of machine vision,
`which may be useful in robotics, inspection or modeling.
`2. Description of the Related Art
`Increasingly more manufacturing operations are per(cid:173)
`formed with the aid of industrial robots. Robots that had
`traditionally been used as blind motion playback machines
`are now benefiting from intelligent sensor-based software to
`adapt to changes in their surroundings. In particular, the use
`of machine vision has been on the rise in industrial robotics.
`A typical vision guided robotic system analyzes image(s)
`from one or more cameras to arrive at such information as
`the position and orientation of a workpiece upon which the
`robotic tool is to operate.
`Early implementations of vision guided robots have pro(cid:173)
`vided only limited part pose information, primarily in the
`two-dimensional space whereby the movement of a given
`part is constrained to a planar surface. For example see U.S.
`Pat. No. 4,437,114 LaRussa. However, many robotic appli(cid:173)
`cations require the robot to locate and manipulate the target
`workpiece in three dimensions. This need has sparked many
`attempts at providing various three-dimensional guidance
`capabilities. In many past cases, this has involved using two 35
`or more cameras that view overlapping regions of the object
`of interest in what is known as a stereo configuration. The
`overlapping images or fields-of-view contain many of the
`same object features viewed from two or more vantage
`points. The difference amongst the apparent position of 40
`corresponding features in each of the images i.e., the par(cid:173)
`allax, is exploited by these methods to calculate the three
`dimensional coordinates of such features. For examples see
`U.S. Pat. No. 4,146,924 Birk et al., and U.S. Pat. No.
`5,959,425 Bieman et al.
`Many drawbacks exist that render stereo based systems
`impractical for industrial applications. The measurement
`error in such systems increases rapidly in response to image
`feature detection errors; these systems also require exactly
`known geometrical relationships between camera pairs. Fur- 50
`thermore stereo methods require the use of at least double
`the number of cameras which drives up the cost, complexity
`and the need for calibration.
`Other attempts at locating objects with multiple cameras
`in the past have taken advantage of video cameras in
`combination with laser light projectors that project various
`stationary or moving patterns such as stripes, cross-hairs and
`the like upon the object of interest. These systems typically
`involve a combination of lasers and cameras that must be
`calibrated relative to a common coordinate system and rely
`on specific assumptions about the geometry of the object of
`interest to work. For example see U.S. Pat. No. 5,160,977
`Utsumi.
`Drawbacks of such attempts include the need for expen(cid:173)
`sive specialized sensors as opposed to use of standard
`off-the-shelf hardware, the need for knowledge of exact
`geometric relationships between all elements of the system
`
`BRIEF SUMMARY OF INVENTION
`
`In one aspect, a method useful in machine-vision of
`objects comprises acquiring a number of images of a first
`view of a training object from a number of cameras; iden(cid:173)
`tifying a number of features of the training object in the
`acquired at least one image of the first view; employing at
`least one of a consistency of physical relationships between
`some of the identified features to set up a system of
`45 equations, where a number of unknowns is not greater than
`a number of equations in the system of equations; and
`automatically computationally solving the system of equa(cid:173)
`tions. The method may further determine a number of
`additional views to be obtained based at least in part on the
`number of image sensors, the number of features identified,
`the number of features having an invariant physical rela-
`tionship associated thereto, and a type of the invariant
`physical relationship associated with the features, sufficient
`to provide a system of equations and unknowns where the
`55 number of unknowns is not greater than the number of
`equations. Where the invariant physical relationships are
`distances, the number of views may, for example, be deter(cid:173)
`mined by computationally solving the equation m~(f2-f-
`2k-2r+6(c-ck))/(f2-3f)-1, where mis the number of views,
`60 f the number of features, k the number of known distances
`between pairs of the features, r is the number of rays with a
`known distance between a feature and an image sensor, c is
`the number of image sensors and ck is the number of known
`transformation between an imager sensor reference frame
`65 and a common reference frame.
`In another aspect, a machine-vision system comprises at
`least one image sensor operable to acquire images of a
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 17 of 29 PageID #: 96
`
`US 7,336,814 B2
`
`3
`trammg object and of target objects; processor-readable
`medium storing instructions for facilitating machine-vision
`for objects having invariant physical relationships between
`a number of features on the objects, by: acquiring a number
`of images of a first view of a training object from a number
`of cameras; identifying a number of features of the training
`object in the acquired at least one image of the first view;
`employing at least one of a consistency of physical relation(cid:173)
`ships between some of the identified features to set up a
`system of equations, where a number of unknowns is not
`greater than a number of equations in the system of equa(cid:173)
`tions; and automatically computationally solving the system
`of equations; and a processor coupled to receive acquired
`images from the at least one image sensor and operable to
`execute the instructions stored in the processor-readable 15
`medium.
`In still another aspect, a processor readable medium stores
`instructions for causing a processor to facilitate machine(cid:173)
`vision for objects having invariant physical relationships
`between a number of features on the objects: by acquiring a 20
`number of images of a first view of a training object from a
`number of cameras; identifying a number of features of the
`training object in the acquired at least one image of the first
`view; employing at least one of a consistency of physical
`relationships between some of the identified features to set 25
`up a system of equations, where a number of unknowns is
`not greater than a number of equations in the system of
`equations; and automatically computationally solving the
`system of equations.
`In a yet another aspect, a method useful in machine-vision 30
`of objects comprises acquiring a number of images of a first
`view of a training object from a number of cameras; iden(cid:173)
`tifying a number of features of the training object in the
`acquired at least one image of the first view; associating
`parameters to less than all of the identified features which 35
`parameters define an invariant physical relationship between
`either the feature and at least one other feature, the feature
`and the at least one camera, or between the at least one
`camera and at least another camera where an invariant
`physical relationship between each one of the features and at 40
`least one other feature is not known when associating the
`parameters before a runtime; determining a number of
`additional views to be obtained based at least in part on the
`number of cameras, the number of features identified, and
`the number of features having parameters associated thereto, 45
`sufficient to provide a system of equations and unknowns
`where the number of unknowns is not greater than the
`number of equations; and acquiring at least one image of
`each of the number of additional views of the training object
`by the at least one camera; identifying at least some of the 50
`number of features of the training object in the acquired at
`least one image of the number of additional views of the
`training object.
`In even another aspect, a method useful in machine-vision
`for objects having invariant physical relationships between 55
`a number of features on the objects comprises in a pre(cid:173)
`runtime environment: acquiring at least one image of a first
`view of a training object by at least one image sensor;
`identifying a number of features of the training object in the
`acquired at least one image of the first view; and associating 60
`a number of parameters to less than all of the identified
`features which define an invariant physical relationship
`between the either the feature and at least one other feature
`or between the feature and the at least one image sensor;
`determining a number of additional views to be obtained
`based at least in part on the number of image sensors
`acquiring at least one image, the number of features of the
`
`4
`training object identified, the number of features having
`parameters associated therewith, and a type of invariant
`physical relationship associated with each of the parameter;
`acquiring at least one image of a second view of the training
`5 object by the at least one image sensor; and identifying at
`least some of the number of features of the training object in
`the acquired at least one image of the second view; and in
`at least one of a pre-run time environment or a runtime
`environment, computationally determining a local model
`10 using the identified features in each of a number of respec(cid:173)
`tive image sensor coordinate frames.
`In still another aspect, a machine-vision system comprises
`at least one image sensor operable to acquire images of a
`training object and of target objects; processor-readable
`medium storing instructions for facilitating pose estimation
`for objects having invariant physical relationships between
`a number of features on the objects, by: in a pre-runtime
`environment: acquiring at least one image of a first view of
`a training object by at least one image sensor; identifying a
`number of features of the training object in the acquired at
`least one image of the first view; and associating a number
`of parameters to less than all of the identified features which
`define an invariant physical relationship between the either
`the feature and at least one other feature or between the
`feature and the at least one image sensor; determining a
`number of additional views to be obtained based at least in
`part on the number of image sensors acquiring at least one
`image, the number of features of the training object identi(cid:173)
`fied, the number of features having parameters associated
`therewith, and a type of invariant physical relationship
`associated with each of the parameter; acquiring at least one
`image of a second view of the training object by the at least
`one image sensor; and identifying at least some of the
`number of features of the training object in the acquired at
`least one image of the second view; and in at least one of a
`pre-run time environment or a runtime environment, com(cid:173)
`putationally determining a local model using the identified
`features in each of a number of respective image sensor
`coordinate frames; and a processor coupled to receive
`acquired images from the at least one image sensor and
`operable to execute the instructions stored in the processor(cid:173)
`readable medium.
`In a further aspect, a processor readable medium stores
`instructions for causing a processor to facilitate machine(cid:173)
`vision for objects having invariant physical relationships
`between a number of features on the objects, by: in a
`pre-runtime environment: acquiring at least one image of a
`first view of a training object by at least one image sensor;
`identifying a number of features of the training object in the
`acquired at least one image of the first view; and associating
`parameters to less than all of the identified features which
`define a physical relationship between the either the feature
`and at least one other feature or between the feature and the
`at least one image sensor; and determining a number of
`additional views to be obtained based at least in part on the
`number of image sensors acquiring at least one image and
`the number of features of the training object identified;
`acquiring at least one image of a second view of the training
`object by the at least one image sensor; and identifying at
`least some of the number of features of the training object in
`the acquired at least one image of the second view; and in
`at least one of a pre-run time environment or a runtime
`65 environment, computationally determining a local model
`using the identified features in each of a number of respec(cid:173)
`tive image sensor coordinate frames.
`
`

`

`Case 1:22-cv-01257-GBW Document 1-2 Filed 09/22/22 Page 18 of 29 PageID #: 97
`
`US 7,336,814 B2
`
`5
`BRIEF DESCRIPTION OF THE SEVERAL
`VIEWS OF THE DRAWING(S)
`
`In the drawings, identical reference numbers identify
`similar elements or acts. The sizes and relative positions of
`elements in the drawings are not necessarily

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket