153:
which are an order of magnitude (10×) smaller. Movement is specifically mentioned here because measuring absolute positions requires an accurate lens model and known reference points within the image to achieve sub-pixel position accuracy. Small movements can however be measured (down to 1 cm) with simple calibration procedures. Specific fit functions often suffer specific bias with respect to image pixel boundaries. Users should therefore take care to avoid these "pixel locking" (or "peak locking") effects.
231:(PSF) of an isolated point in the image. If the image does not contain isolated points, similar methods can be applied to edges in the image. It is also important when attempting sub-pixel resolution to keep image noise to a minimum. This, in the case of a stationary scene, can be measured from a time series of images. Appropriate pixel averaging, through both time (for stationary images) and space (for uniform regions of the image) is often used to prepare the image for sub-pixel resolution measurements.
122:
167:
37:
152:
For example, if the image of a ship of length 50 metres (160 ft), viewed side-on, is 500 pixels long, the nominal resolution (pixel size) on the side of the ship facing the camera is 0.1 metres (3.9 in). Now sub-pixel resolution of well resolved features can measure ship movements
188:
58:
17:
354:
Psarakis, E. Z.; Evangelidis, G. D. (2005). "An enhanced correlation-based method for stereo correspondence with subpixel accuracy".
376:
328:
214:
84:
192:
62:
311:
Nehab, D.; Rusinkiewiez, S.; Davis, J. (2005). "Improved sub-pixel stereo correspondences through symmetric refinement".
227:
Whether features in a digital image are sharp enough to achieve sub-pixel resolution can be quantified by measuring the
412:
355:
177:
98:
47:
274:
Shimizu, M.; Okutomi, M. (2003). "Significance and attributes of subpixel estimation on area-based matching".
196:
181:
66:
51:
228:
390:
342:
299:
25:
382:
372:
334:
324:
291:
364:
316:
283:
106:
105:
can be obtained in images constructed from sources with information exceeding the nominal
121:
406:
346:
303:
394:
166:
36:
16:
386:
338:
295:
320:
368:
357:
Tenth IEEE International
Conference on Computer Vision (ICCV'05) Volume 1
313:
Tenth IEEE International
Conference on Computer Vision (ICCV'05) Volume 1
287:
248:
15:
160:
116:
30:
133:
249:Accurate particle position measurement from images
257:Virtual Journal of Biological Physics Research
8:
255:, Vol. 78, 053704 (2007); also selected for
195:. Unsourced material may be challenged and
65:. Unsourced material may be challenged and
215:Learn how and when to remove this message
85:Learn how and when to remove this message
240:
7:
193:adding citations to reliable sources
63:adding citations to reliable sources
251:". Y. Feng, J. Goree, and Bin Liu,
14:
253:Review of Scientific Instruments
165:
120:
35:
20:Sub-pixel rendering of a circle
276:Systems and Computers in Japan
1:
429:
23:
99:digital image processing
24:Not to be confused with
157:Determining feasibility
21:
321:10.1109/ICCV.2005.119
229:point spread function
19:
369:10.1109/ICCV.2005.33
363:. pp. 907–912.
315:. pp. 557–563.
189:improve this section
103:sub-pixel resolution
59:improve this section
259:, Vol. 13, (2007).
132:. You can help by
26:Subpixel rendering
22:
288:10.1002/scj.10506
225:
224:
217:
150:
149:
109:of said images.
95:
94:
87:
420:
413:Image processing
398:
362:
350:
307:
260:
245:
220:
213:
209:
206:
200:
169:
161:
145:
142:
124:
117:
107:pixel resolution
90:
83:
79:
76:
70:
39:
31:
428:
427:
423:
422:
421:
419:
418:
417:
403:
402:
401:
379:
360:
353:
331:
310:
273:
269:
264:
263:
246:
242:
237:
221:
210:
204:
201:
186:
170:
159:
146:
140:
137:
130:needs expansion
115:
91:
80:
74:
71:
56:
40:
29:
12:
11:
5:
426:
424:
416:
415:
405:
404:
400:
399:
377:
351:
329:
308:
270:
268:
265:
262:
261:
239:
238:
236:
233:
223:
222:
173:
171:
164:
158:
155:
148:
147:
127:
125:
114:
111:
93:
92:
43:
41:
34:
13:
10:
9:
6:
4:
3:
2:
425:
414:
411:
410:
408:
396:
392:
388:
384:
380:
378:0-7695-2334-X
374:
370:
366:
359:
358:
352:
348:
344:
340:
336:
332:
330:0-7695-2334-X
326:
322:
318:
314:
309:
305:
301:
297:
293:
289:
285:
282:(12): 1–111.
281:
277:
272:
271:
266:
258:
254:
250:
244:
241:
234:
232:
230:
219:
216:
208:
198:
194:
190:
184:
183:
179:
174:This section
172:
168:
163:
162:
156:
154:
144:
135:
131:
128:This section
126:
123:
119:
118:
112:
110:
108:
104:
100:
89:
86:
78:
68:
64:
60:
54:
53:
49:
44:This section
42:
38:
33:
32:
27:
18:
356:
312:
279:
275:
256:
252:
243:
226:
211:
202:
187:Please help
175:
151:
141:January 2011
138:
134:adding to it
129:
102:
96:
81:
72:
57:Please help
45:
267:References
205:March 2024
75:March 2024
387:1550-5499
339:1550-5499
296:1520-684X
235:Footnotes
176:does not
46:does not
407:Category
347:14172959
304:41202105
395:2723727
197:removed
182:sources
113:Example
67:removed
52:sources
393:
385:
375:
345:
337:
327:
302:
294:
391:S2CID
361:(PDF)
343:S2CID
300:S2CID
383:ISSN
373:ISBN
335:ISSN
325:ISBN
292:ISSN
180:any
178:cite
50:any
48:cite
365:doi
317:doi
284:doi
191:by
136:.
97:In
61:by
409::
389:.
381:.
371:.
341:.
333:.
323:.
298:.
290:.
280:34
278:.
101:,
397:.
367::
349:.
319::
306:.
286::
247:"
218:)
212:(
207:)
203:(
199:.
185:.
143:)
139:(
88:)
82:(
77:)
73:(
69:.
55:.
28:.
Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.