22:
158:
commercial data packages also commonly include both lower-resolution multispectral bands and a single panchromatic band. One of the principal reasons for configuring satellite sensors this way is to keep satellite weight, cost, bandwidth and complexity down. Pan sharpening uses spatial information
200:
Pan-sharpening techniques can result in spectral distortions when pan sharpening satellite images as a result of the nature of the panchromatic band. The
Landsat panchromatic band for example is not sensitive to blue light. As a result, the spectral characteristics of the raw pansharpened color
133:
and nearly every map creating company use this technique to increase image quality. Pansharpening produces a high-resolution color image from three, four or more low-resolution multispectral satellite bands plus a corresponding high-resolution panchromatic band:
201:
image may not exactly match those of the corresponding low-resolution RGB image, resulting in altered color tones. This has resulted in the development of many algorithms that attempt to reduce this spectral distortion and to produce visually pleasing images.
159:
in the high-resolution grayscale band and color information in the multispectral bands to create a high-resolution color image, essentially increasing the resolution of the color information in the data set to match that of the panchromatic band.
169:
Alignment: the up-sampled color bands and the panchromatic band are aligned to reduce artifacts due to mis-registration (generally, when the data comes from the same sensor, this step is usually not necessary);
181:
Reverse transform: the reverse transformation is performed using the substituted intensity component to transform back to the original color space.
172:
Forward transform: the up-sampled color bands are transformed to an alternate color space (where intensity is orthogonal to the color information);
146:, which includes six 30 m resolution multispectral bands, a 60 m thermal infrared band plus a 15 m resolution panchromatic band.
105:
162:
One common class of algorithms for pansharpening is called “component substitution,” which usually involves the following steps:
312:
43:
39:
86:
58:
194:
268:
65:
175:
Intensity matching: the intensity of the color bands is matched to the pan band intensity in the transformed space;
32:
72:
317:
178:
Component substitution: the pan band is then directly substituted for the transformed intensity component;
54:
279:
155:
322:
248:
166:
Up-sampling: the color bands are up-sampled to the same resolution as the panchromatic band;
147:
79:
292:
327:
306:
126:
215:
122:
210:
186:
142:
Such band combinations are commonly bundled in satellite data sets, for example
130:
21:
143:
253:
236:
193:. The same steps can also be performed using wavelet decomposition or
151:
138:
Low-res color bands + High-res grayscale band = High-res color image
190:
237:"Smart pansharpening approach using kernel-based image filtering"
185:
Common color-space transformation used for pan sharpening are
15:
129:
imagery to create a single high-resolution color image.
197:
and replacing the first component with the pan band.
46:. Unsourced material may be challenged and removed.
136:
8:
252:
106:Learn how and when to remove this message
121:is a process of merging high-resolution
227:
288:
277:
7:
44:adding citations to reliable sources
14:
189:(hue-saturation-intensity), and
20:
235:AL Smadi, Ahmad (18 May 2021).
31:needs additional citations for
1:
269:"WorldView-2 Pan-Sharpening"
344:
313:Photographic processes
287:Cite journal requires
140:
125:and lower resolution
241:IET Image Processing
40:improve this article
254:10.1049/ipr2.12251
247:(11): 2629–2642.
116:
115:
108:
90:
335:
297:
296:
290:
285:
283:
275:
273:
265:
259:
258:
256:
232:
111:
104:
100:
97:
91:
89:
48:
24:
16:
343:
342:
338:
337:
336:
334:
333:
332:
303:
302:
301:
300:
286:
276:
271:
267:
266:
262:
234:
233:
229:
224:
207:
112:
101:
95:
92:
55:"Pansharpening"
49:
47:
37:
25:
12:
11:
5:
341:
339:
331:
330:
325:
320:
318:Remote sensing
315:
305:
304:
299:
298:
289:|journal=
260:
226:
225:
223:
220:
219:
218:
213:
206:
203:
183:
182:
179:
176:
173:
170:
167:
114:
113:
96:September 2014
28:
26:
19:
13:
10:
9:
6:
4:
3:
2:
340:
329:
326:
324:
321:
319:
316:
314:
311:
310:
308:
294:
281:
270:
264:
261:
255:
250:
246:
242:
238:
231:
228:
221:
217:
214:
212:
209:
208:
204:
202:
198:
196:
192:
188:
180:
177:
174:
171:
168:
165:
164:
163:
160:
157:
153:
149:
145:
139:
135:
132:
128:
127:multispectral
124:
120:
119:Pansharpening
110:
107:
99:
88:
85:
81:
78:
74:
71:
67:
64:
60:
57: –
56:
52:
51:Find sources:
45:
41:
35:
34:
29:This article
27:
23:
18:
17:
280:cite journal
263:
244:
240:
230:
216:Image fusion
199:
184:
161:
141:
137:
123:panchromatic
118:
117:
102:
93:
83:
76:
69:
62:
50:
38:Please help
33:verification
30:
211:Demosaicing
131:Google Maps
307:Categories
222:References
66:newspapers
144:Landsat 7
205:See also
323:Imaging
80:scholar
152:GeoEye
82:
75:
68:
61:
53:
328:Color
272:(PDF)
191:YCbCr
156:Maxar
87:JSTOR
73:books
293:help
154:and
148:SPOT
59:news
249:doi
195:PCA
187:HSI
42:by
309::
284::
282:}}
278:{{
245:15
243:.
239:.
150:,
295:)
291:(
274:.
257:.
251::
109:)
103:(
98:)
94:(
84:·
77:·
70:·
63:·
36:.
Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.