A new automatic method capable of registering multimodal images like terrain maps and multispectral data is presented. In order to speed up the processing, given the large amount of data typical of such settings, the method exploits a multi-resolution approach, which may select different similarity measures in consideration of image resolution and size. In fact, the performances of Cross Correlation and Maximization of Mutual Information (MMI) on images of different resolution and size have been evaluated and are described The adaptive strategy adopted is designed to exploit the strengths and to overcome the limitations of the similarity criteria employed. In case multimodal views are to be registered on 3D models, MMI is to be preferred. The strategies to improve its performance also on smaller images are presented.

Automatic registration of multimodal views on large aerial images

F. Uccheddu;
2012

Abstract

A new automatic method capable of registering multimodal images like terrain maps and multispectral data is presented. In order to speed up the processing, given the large amount of data typical of such settings, the method exploits a multi-resolution approach, which may select different similarity measures in consideration of image resolution and size. In fact, the performances of Cross Correlation and Maximization of Mutual Information (MMI) on images of different resolution and size have been evaluated and are described The adaptive strategy adopted is designed to exploit the strengths and to overcome the limitations of the similarity criteria employed. In case multimodal views are to be registered on 3D models, MMI is to be preferred. The strategies to improve its performance also on smaller images are presented.
2012
Proc. of SPIE, 8537
Image and Signal Processing for Remote Sensing XVIII
File in questo prodotto:
Non ci sono file associati a questo prodotto.
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11577/3379880
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
  • OpenAlex ND
social impact