Knowledge (XXG)

Pose (computer vision)

Source đź“ť

87:
of the object's pose. Once a set of control points on the object, typically corners or other feature points, has been identified, it is then possible to solve the pose transformation from a set of equations which relate the 3D coordinates of the points with their 2D image coordinates. Algorithms that determine the pose of a
116:
Learning-based methods: These methods use artificial learning-based system which learn the mapping from 2D image features to pose transformation. In short, this means that a sufficiently large set of images of the object, in different poses, must be presented to the system during a learning phase.
86:
Analytic or geometric methods: Given that the image sensor (camera) is calibrated and the mapping from 3D points in the scene and 2D points in the image is known. If also the geometry of the object is known, it means that the projected image of the object on the camera image is a well-known function
64:. This information can then be used, for example, to allow a robot to manipulate an object or to avoid moving into the object based on its perceived position and orientation in the environment. Other applications include skeletal action recognition. 174:
or be restricted for the estimation of the intrinsic parameters only. Exterior orientation and interior orientation refer to the determination of only the extrinsic and intrinsic parameters, respectively.
246: 82:. Pose estimation problems can be solved in different ways depending on the image sensor configuration, and choice of methodology. Three classes of methodologies can be distinguished: 105:
may be used. This approach is robust especially when the images are not perfectly calibrated. In this particular case, the pose represent the
154: 359: 117:
Once the learning phase is completed, the system should be able to present an estimate of the object's pose given an image of the object.
292: 167:(position and orientation) while the intrinsic parameters specify the camera image format (focal length, pixel size, and image origin). 190:
where the camera projection matrices of two cameras are used to calculate the 3D world coordinates of a point viewed by both cameras.
171: 170:
This process is often called geometric camera calibration or simply camera calibration, although that term may also refer to
306: 205: 78:
The specific task of determining the pose of an object in an image (or stereo images, image sequence) is referred to as
354: 180: 364: 106: 92: 40: 320: 251:
Intelligent Robots and Computer Vision XV: Algorithms, Techniques, Active Vision, and Materials Handling
215: 48: 258: 139: 51:. The term “pose” is largely synonymous with the term “transform”, but a transform may often include 210: 200: 135: 128: 36: 334: 274: 142:
approximating the camera that produced a given photograph or video; it determines which incoming
73: 58:
In computer vision, the pose of an object is often estimated from camera input by the process of
52: 178:
The classic camera calibration requires special objects in the scene, which is not required in
102: 98: 266: 220: 147: 146:
is associated with each pixel on the resulting image. Basically, the process determines the
110: 44: 224: 60: 24: 307:"Transformation matrices to geometry_msgs/Pose - ROS Answers: Open Source Q&A Forum" 262: 109:
and the error between the projection of the object control points with the image is the
245:
Hoff, William A.; Nguyen, Khoi; Lyon, Torsten (1996-10-29). Casasent, David P. (ed.).
348: 187: 159: 278: 88: 101:
methods: If the pose of an object does not have to be computed in real-time a
143: 20: 153:
Usually, the camera parameters are represented in a 3 × 4
95:
algorithms, if the correspondences between points are not already known.
270: 247:"Computer-vision-based registration techniques for augmented reality" 186:
Camera resectioning is often used in the application of
138:is the process of estimating the parameters of a 91:with respect to another point cloud are known as 16:Position and orientation of an object in an image 163:. The extrinsic parameters define the camera 8: 237: 47:. Poses are often stored internally as 7: 321:"Drake: Spatial Pose and Transform" 14: 293:"Pose (Position and Orientation)" 127:This section is an excerpt from 335:"Apple Developer Documentation" 172:photometric camera calibration 1: 206:Homography (computer vision) 360:Geometry in computer vision 381: 126: 71: 55:, whereas pose does not. 43:of an object, usually in 181:camera auto-calibration 150:of the pinhole camera. 49:transformation matrices 107:genetic representation 93:point set registration 216:Structure from motion 140:pinhole camera model 263:1996SPIE.2904..538H 201:Gesture recognition 136:Camera resectioning 129:Camera resectioning 211:Camera calibration 74:3D pose estimation 271:10.1117/12.256311 257:. SPIE: 538–548. 155:projection matrix 103:genetic algorithm 99:Genetic algorithm 35:) represents the 19:In the fields of 372: 339: 338: 331: 325: 324: 317: 311: 310: 303: 297: 296: 289: 283: 282: 242: 221:Essential matrix 111:fitness function 45:three dimensions 380: 379: 375: 374: 373: 371: 370: 369: 355:Computer vision 345: 344: 343: 342: 333: 332: 328: 319: 318: 314: 305: 304: 300: 291: 290: 286: 244: 243: 239: 234: 227:(relative pose) 225:Trifocal tensor 197: 192: 191: 132: 124: 80:pose estimation 76: 70: 68:Pose estimation 61:pose estimation 25:computer vision 17: 12: 11: 5: 378: 376: 368: 367: 362: 357: 347: 346: 341: 340: 326: 312: 298: 284: 236: 235: 233: 230: 229: 228: 218: 213: 208: 203: 196: 193: 133: 125: 123: 120: 119: 118: 114: 96: 72:Main article: 69: 66: 15: 13: 10: 9: 6: 4: 3: 2: 377: 366: 365:Robot control 363: 361: 358: 356: 353: 352: 350: 336: 330: 327: 322: 316: 313: 308: 302: 299: 294: 288: 285: 280: 276: 272: 268: 264: 260: 256: 252: 248: 241: 238: 231: 226: 222: 219: 217: 214: 212: 209: 207: 204: 202: 199: 198: 194: 189: 188:stereo vision 185: 183: 182: 176: 173: 168: 166: 162: 161: 160:camera matrix 156: 151: 149: 145: 141: 137: 130: 121: 115: 112: 108: 104: 100: 97: 94: 90: 85: 84: 83: 81: 75: 67: 65: 63: 62: 56: 54: 50: 46: 42: 38: 34: 30: 26: 22: 329: 315: 301: 287: 254: 250: 240: 179: 177: 169: 164: 158: 152: 134: 79: 77: 59: 57: 33:spatial pose 32: 28: 18: 157:called the 122:Camera pose 89:point cloud 41:orientation 349:Categories 232:References 144:light ray 21:computing 195:See also 37:position 279:6587175 259:Bibcode 277:  275:S2CID 53:scale 255:2904 223:and 165:pose 148:pose 39:and 31:(or 29:pose 23:and 267:doi 351:: 273:. 265:. 253:. 249:. 184:. 27:, 337:. 323:. 309:. 295:. 281:. 269:: 261:: 131:. 113:.

Index

computing
computer vision
position
orientation
three dimensions
transformation matrices
scale
pose estimation
3D pose estimation
point cloud
point set registration
Genetic algorithm
genetic algorithm
genetic representation
fitness function
Camera resectioning
Camera resectioning
pinhole camera model
light ray
pose
projection matrix
camera matrix
photometric camera calibration
camera auto-calibration
stereo vision
Gesture recognition
Homography (computer vision)
Camera calibration
Structure from motion
Essential matrix

Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.

↑