An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion
Abstract Infrared and visible image fusion aims to obtain a more informative fusion image by merging the infrared and visible images. However, the existing methods have some shortcomings, such as detail information loss, unclear boundaries, and not being end-to-end. In this paper, we propose an end-...
Ausführliche Beschreibung
Autor*in: |
Liu, Hongzhe [verfasserIn] |
---|
Format: |
Artikel |
---|---|
Sprache: |
Englisch |
Erschienen: |
2022 |
---|
Schlagwörter: |
---|
Anmerkung: |
© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. |
---|
Übergeordnetes Werk: |
Enthalten in: Multimedia tools and applications - Springer US, 1995, 82(2022), 13 vom: 27. Dez., Seite 20139-20156 |
---|---|
Übergeordnetes Werk: |
volume:82 ; year:2022 ; number:13 ; day:27 ; month:12 ; pages:20139-20156 |
Links: |
---|
DOI / URN: |
10.1007/s11042-022-14314-9 |
---|
Katalog-ID: |
OLC2134624426 |
---|
LEADER | 01000naa a22002652 4500 | ||
---|---|---|---|
001 | OLC2134624426 | ||
003 | DE-627 | ||
005 | 20230510162757.0 | ||
007 | tu | ||
008 | 230510s2022 xx ||||| 00| ||eng c | ||
024 | 7 | |a 10.1007/s11042-022-14314-9 |2 doi | |
035 | |a (DE-627)OLC2134624426 | ||
035 | |a (DE-He213)s11042-022-14314-9-p | ||
040 | |a DE-627 |b ger |c DE-627 |e rakwb | ||
041 | |a eng | ||
082 | 0 | 4 | |a 070 |a 004 |q VZ |
100 | 1 | |a Liu, Hongzhe |e verfasserin |4 aut | |
245 | 1 | 0 | |a An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion |
264 | 1 | |c 2022 | |
336 | |a Text |b txt |2 rdacontent | ||
337 | |a ohne Hilfsmittel zu benutzen |b n |2 rdamedia | ||
338 | |a Band |b nc |2 rdacarrier | ||
500 | |a © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. | ||
520 | |a Abstract Infrared and visible image fusion aims to obtain a more informative fusion image by merging the infrared and visible images. However, the existing methods have some shortcomings, such as detail information loss, unclear boundaries, and not being end-to-end. In this paper, we propose an end-to-end network architecture for infrared and visible image fusion task. Our network contains three essential parts: encoders, residual fusion module, and decoder. First, we input infrared and visible images to two encoders to extract shallow features, respectively. Subsequently, the two sets of features are concatenated and fed to the residual fusion module to extract multi-scale features and fuse them adequately. Finally, the fused image is obtained by the decoder. We conduct objective and subjective experiments on two public datasets. The comparison results with the state-of-art methods prove that the fusion results of the proposed method have better objective metrics and contain more detail information and more explicit boundary. | ||
650 | 4 | |a Autoencoder | |
650 | 4 | |a Image fusion | |
650 | 4 | |a Infrared image | |
650 | 4 | |a Visible image | |
700 | 1 | |a Yan, Hua |0 (orcid)0000-0001-9231-3175 |4 aut | |
773 | 0 | 8 | |i Enthalten in |t Multimedia tools and applications |d Springer US, 1995 |g 82(2022), 13 vom: 27. Dez., Seite 20139-20156 |w (DE-627)189064145 |w (DE-600)1287642-2 |w (DE-576)052842126 |x 1380-7501 |7 nnns |
773 | 1 | 8 | |g volume:82 |g year:2022 |g number:13 |g day:27 |g month:12 |g pages:20139-20156 |
856 | 4 | 1 | |u https://doi.org/10.1007/s11042-022-14314-9 |z lizenzpflichtig |3 Volltext |
912 | |a GBV_USEFLAG_A | ||
912 | |a SYSFLAG_A | ||
912 | |a GBV_OLC | ||
912 | |a SSG-OLC-MAT | ||
912 | |a SSG-OLC-BUB | ||
912 | |a SSG-OLC-MKW | ||
951 | |a AR | ||
952 | |d 82 |j 2022 |e 13 |b 27 |c 12 |h 20139-20156 |
author_variant |
h l hl h y hy |
---|---|
matchkey_str |
article:13807501:2022----::nntedutsaeewrbsdnuonoefrnrrd |
hierarchy_sort_str |
2022 |
publishDate |
2022 |
allfields |
10.1007/s11042-022-14314-9 doi (DE-627)OLC2134624426 (DE-He213)s11042-022-14314-9-p DE-627 ger DE-627 rakwb eng 070 004 VZ Liu, Hongzhe verfasserin aut An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion 2022 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. Abstract Infrared and visible image fusion aims to obtain a more informative fusion image by merging the infrared and visible images. However, the existing methods have some shortcomings, such as detail information loss, unclear boundaries, and not being end-to-end. In this paper, we propose an end-to-end network architecture for infrared and visible image fusion task. Our network contains three essential parts: encoders, residual fusion module, and decoder. First, we input infrared and visible images to two encoders to extract shallow features, respectively. Subsequently, the two sets of features are concatenated and fed to the residual fusion module to extract multi-scale features and fuse them adequately. Finally, the fused image is obtained by the decoder. We conduct objective and subjective experiments on two public datasets. The comparison results with the state-of-art methods prove that the fusion results of the proposed method have better objective metrics and contain more detail information and more explicit boundary. Autoencoder Image fusion Infrared image Visible image Yan, Hua (orcid)0000-0001-9231-3175 aut Enthalten in Multimedia tools and applications Springer US, 1995 82(2022), 13 vom: 27. Dez., Seite 20139-20156 (DE-627)189064145 (DE-600)1287642-2 (DE-576)052842126 1380-7501 nnns volume:82 year:2022 number:13 day:27 month:12 pages:20139-20156 https://doi.org/10.1007/s11042-022-14314-9 lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT SSG-OLC-BUB SSG-OLC-MKW AR 82 2022 13 27 12 20139-20156 |
spelling |
10.1007/s11042-022-14314-9 doi (DE-627)OLC2134624426 (DE-He213)s11042-022-14314-9-p DE-627 ger DE-627 rakwb eng 070 004 VZ Liu, Hongzhe verfasserin aut An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion 2022 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. Abstract Infrared and visible image fusion aims to obtain a more informative fusion image by merging the infrared and visible images. However, the existing methods have some shortcomings, such as detail information loss, unclear boundaries, and not being end-to-end. In this paper, we propose an end-to-end network architecture for infrared and visible image fusion task. Our network contains three essential parts: encoders, residual fusion module, and decoder. First, we input infrared and visible images to two encoders to extract shallow features, respectively. Subsequently, the two sets of features are concatenated and fed to the residual fusion module to extract multi-scale features and fuse them adequately. Finally, the fused image is obtained by the decoder. We conduct objective and subjective experiments on two public datasets. The comparison results with the state-of-art methods prove that the fusion results of the proposed method have better objective metrics and contain more detail information and more explicit boundary. Autoencoder Image fusion Infrared image Visible image Yan, Hua (orcid)0000-0001-9231-3175 aut Enthalten in Multimedia tools and applications Springer US, 1995 82(2022), 13 vom: 27. Dez., Seite 20139-20156 (DE-627)189064145 (DE-600)1287642-2 (DE-576)052842126 1380-7501 nnns volume:82 year:2022 number:13 day:27 month:12 pages:20139-20156 https://doi.org/10.1007/s11042-022-14314-9 lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT SSG-OLC-BUB SSG-OLC-MKW AR 82 2022 13 27 12 20139-20156 |
allfields_unstemmed |
10.1007/s11042-022-14314-9 doi (DE-627)OLC2134624426 (DE-He213)s11042-022-14314-9-p DE-627 ger DE-627 rakwb eng 070 004 VZ Liu, Hongzhe verfasserin aut An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion 2022 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. Abstract Infrared and visible image fusion aims to obtain a more informative fusion image by merging the infrared and visible images. However, the existing methods have some shortcomings, such as detail information loss, unclear boundaries, and not being end-to-end. In this paper, we propose an end-to-end network architecture for infrared and visible image fusion task. Our network contains three essential parts: encoders, residual fusion module, and decoder. First, we input infrared and visible images to two encoders to extract shallow features, respectively. Subsequently, the two sets of features are concatenated and fed to the residual fusion module to extract multi-scale features and fuse them adequately. Finally, the fused image is obtained by the decoder. We conduct objective and subjective experiments on two public datasets. The comparison results with the state-of-art methods prove that the fusion results of the proposed method have better objective metrics and contain more detail information and more explicit boundary. Autoencoder Image fusion Infrared image Visible image Yan, Hua (orcid)0000-0001-9231-3175 aut Enthalten in Multimedia tools and applications Springer US, 1995 82(2022), 13 vom: 27. Dez., Seite 20139-20156 (DE-627)189064145 (DE-600)1287642-2 (DE-576)052842126 1380-7501 nnns volume:82 year:2022 number:13 day:27 month:12 pages:20139-20156 https://doi.org/10.1007/s11042-022-14314-9 lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT SSG-OLC-BUB SSG-OLC-MKW AR 82 2022 13 27 12 20139-20156 |
allfieldsGer |
10.1007/s11042-022-14314-9 doi (DE-627)OLC2134624426 (DE-He213)s11042-022-14314-9-p DE-627 ger DE-627 rakwb eng 070 004 VZ Liu, Hongzhe verfasserin aut An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion 2022 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. Abstract Infrared and visible image fusion aims to obtain a more informative fusion image by merging the infrared and visible images. However, the existing methods have some shortcomings, such as detail information loss, unclear boundaries, and not being end-to-end. In this paper, we propose an end-to-end network architecture for infrared and visible image fusion task. Our network contains three essential parts: encoders, residual fusion module, and decoder. First, we input infrared and visible images to two encoders to extract shallow features, respectively. Subsequently, the two sets of features are concatenated and fed to the residual fusion module to extract multi-scale features and fuse them adequately. Finally, the fused image is obtained by the decoder. We conduct objective and subjective experiments on two public datasets. The comparison results with the state-of-art methods prove that the fusion results of the proposed method have better objective metrics and contain more detail information and more explicit boundary. Autoencoder Image fusion Infrared image Visible image Yan, Hua (orcid)0000-0001-9231-3175 aut Enthalten in Multimedia tools and applications Springer US, 1995 82(2022), 13 vom: 27. Dez., Seite 20139-20156 (DE-627)189064145 (DE-600)1287642-2 (DE-576)052842126 1380-7501 nnns volume:82 year:2022 number:13 day:27 month:12 pages:20139-20156 https://doi.org/10.1007/s11042-022-14314-9 lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT SSG-OLC-BUB SSG-OLC-MKW AR 82 2022 13 27 12 20139-20156 |
allfieldsSound |
10.1007/s11042-022-14314-9 doi (DE-627)OLC2134624426 (DE-He213)s11042-022-14314-9-p DE-627 ger DE-627 rakwb eng 070 004 VZ Liu, Hongzhe verfasserin aut An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion 2022 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. Abstract Infrared and visible image fusion aims to obtain a more informative fusion image by merging the infrared and visible images. However, the existing methods have some shortcomings, such as detail information loss, unclear boundaries, and not being end-to-end. In this paper, we propose an end-to-end network architecture for infrared and visible image fusion task. Our network contains three essential parts: encoders, residual fusion module, and decoder. First, we input infrared and visible images to two encoders to extract shallow features, respectively. Subsequently, the two sets of features are concatenated and fed to the residual fusion module to extract multi-scale features and fuse them adequately. Finally, the fused image is obtained by the decoder. We conduct objective and subjective experiments on two public datasets. The comparison results with the state-of-art methods prove that the fusion results of the proposed method have better objective metrics and contain more detail information and more explicit boundary. Autoencoder Image fusion Infrared image Visible image Yan, Hua (orcid)0000-0001-9231-3175 aut Enthalten in Multimedia tools and applications Springer US, 1995 82(2022), 13 vom: 27. Dez., Seite 20139-20156 (DE-627)189064145 (DE-600)1287642-2 (DE-576)052842126 1380-7501 nnns volume:82 year:2022 number:13 day:27 month:12 pages:20139-20156 https://doi.org/10.1007/s11042-022-14314-9 lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT SSG-OLC-BUB SSG-OLC-MKW AR 82 2022 13 27 12 20139-20156 |
language |
English |
source |
Enthalten in Multimedia tools and applications 82(2022), 13 vom: 27. Dez., Seite 20139-20156 volume:82 year:2022 number:13 day:27 month:12 pages:20139-20156 |
sourceStr |
Enthalten in Multimedia tools and applications 82(2022), 13 vom: 27. Dez., Seite 20139-20156 volume:82 year:2022 number:13 day:27 month:12 pages:20139-20156 |
format_phy_str_mv |
Article |
institution |
findex.gbv.de |
topic_facet |
Autoencoder Image fusion Infrared image Visible image |
dewey-raw |
070 |
isfreeaccess_bool |
false |
container_title |
Multimedia tools and applications |
authorswithroles_txt_mv |
Liu, Hongzhe @@aut@@ Yan, Hua @@aut@@ |
publishDateDaySort_date |
2022-12-27T00:00:00Z |
hierarchy_top_id |
189064145 |
dewey-sort |
270 |
id |
OLC2134624426 |
language_de |
englisch |
fullrecord |
<?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01000naa a22002652 4500</leader><controlfield tag="001">OLC2134624426</controlfield><controlfield tag="003">DE-627</controlfield><controlfield tag="005">20230510162757.0</controlfield><controlfield tag="007">tu</controlfield><controlfield tag="008">230510s2022 xx ||||| 00| ||eng c</controlfield><datafield tag="024" ind1="7" ind2=" "><subfield code="a">10.1007/s11042-022-14314-9</subfield><subfield code="2">doi</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-627)OLC2134624426</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-He213)s11042-022-14314-9-p</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-627</subfield><subfield code="b">ger</subfield><subfield code="c">DE-627</subfield><subfield code="e">rakwb</subfield></datafield><datafield tag="041" ind1=" " ind2=" "><subfield code="a">eng</subfield></datafield><datafield tag="082" ind1="0" ind2="4"><subfield code="a">070</subfield><subfield code="a">004</subfield><subfield code="q">VZ</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Liu, Hongzhe</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="c">2022</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">Text</subfield><subfield code="b">txt</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">ohne Hilfsmittel zu benutzen</subfield><subfield code="b">n</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">Band</subfield><subfield code="b">nc</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a">© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">Abstract Infrared and visible image fusion aims to obtain a more informative fusion image by merging the infrared and visible images. However, the existing methods have some shortcomings, such as detail information loss, unclear boundaries, and not being end-to-end. In this paper, we propose an end-to-end network architecture for infrared and visible image fusion task. Our network contains three essential parts: encoders, residual fusion module, and decoder. First, we input infrared and visible images to two encoders to extract shallow features, respectively. Subsequently, the two sets of features are concatenated and fed to the residual fusion module to extract multi-scale features and fuse them adequately. Finally, the fused image is obtained by the decoder. We conduct objective and subjective experiments on two public datasets. The comparison results with the state-of-art methods prove that the fusion results of the proposed method have better objective metrics and contain more detail information and more explicit boundary.</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Autoencoder</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Image fusion</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Infrared image</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Visible image</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Yan, Hua</subfield><subfield code="0">(orcid)0000-0001-9231-3175</subfield><subfield code="4">aut</subfield></datafield><datafield tag="773" ind1="0" ind2="8"><subfield code="i">Enthalten in</subfield><subfield code="t">Multimedia tools and applications</subfield><subfield code="d">Springer US, 1995</subfield><subfield code="g">82(2022), 13 vom: 27. Dez., Seite 20139-20156</subfield><subfield code="w">(DE-627)189064145</subfield><subfield code="w">(DE-600)1287642-2</subfield><subfield code="w">(DE-576)052842126</subfield><subfield code="x">1380-7501</subfield><subfield code="7">nnns</subfield></datafield><datafield tag="773" ind1="1" ind2="8"><subfield code="g">volume:82</subfield><subfield code="g">year:2022</subfield><subfield code="g">number:13</subfield><subfield code="g">day:27</subfield><subfield code="g">month:12</subfield><subfield code="g">pages:20139-20156</subfield></datafield><datafield tag="856" ind1="4" ind2="1"><subfield code="u">https://doi.org/10.1007/s11042-022-14314-9</subfield><subfield code="z">lizenzpflichtig</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_USEFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SYSFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_OLC</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-MAT</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-BUB</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-MKW</subfield></datafield><datafield tag="951" ind1=" " ind2=" "><subfield code="a">AR</subfield></datafield><datafield tag="952" ind1=" " ind2=" "><subfield code="d">82</subfield><subfield code="j">2022</subfield><subfield code="e">13</subfield><subfield code="b">27</subfield><subfield code="c">12</subfield><subfield code="h">20139-20156</subfield></datafield></record></collection>
|
author |
Liu, Hongzhe |
spellingShingle |
Liu, Hongzhe ddc 070 misc Autoencoder misc Image fusion misc Infrared image misc Visible image An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion |
authorStr |
Liu, Hongzhe |
ppnlink_with_tag_str_mv |
@@773@@(DE-627)189064145 |
format |
Article |
dewey-ones |
070 - News media, journalism & publishing 004 - Data processing & computer science |
delete_txt_mv |
keep |
author_role |
aut aut |
collection |
OLC |
remote_str |
false |
illustrated |
Not Illustrated |
issn |
1380-7501 |
topic_title |
070 004 VZ An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion Autoencoder Image fusion Infrared image Visible image |
topic |
ddc 070 misc Autoencoder misc Image fusion misc Infrared image misc Visible image |
topic_unstemmed |
ddc 070 misc Autoencoder misc Image fusion misc Infrared image misc Visible image |
topic_browse |
ddc 070 misc Autoencoder misc Image fusion misc Infrared image misc Visible image |
format_facet |
Aufsätze Gedruckte Aufsätze |
format_main_str_mv |
Text Zeitschrift/Artikel |
carriertype_str_mv |
nc |
hierarchy_parent_title |
Multimedia tools and applications |
hierarchy_parent_id |
189064145 |
dewey-tens |
070 - News media, journalism & publishing 000 - Computer science, knowledge & systems |
hierarchy_top_title |
Multimedia tools and applications |
isfreeaccess_txt |
false |
familylinks_str_mv |
(DE-627)189064145 (DE-600)1287642-2 (DE-576)052842126 |
title |
An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion |
ctrlnum |
(DE-627)OLC2134624426 (DE-He213)s11042-022-14314-9-p |
title_full |
An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion |
author_sort |
Liu, Hongzhe |
journal |
Multimedia tools and applications |
journalStr |
Multimedia tools and applications |
lang_code |
eng |
isOA_bool |
false |
dewey-hundreds |
000 - Computer science, information & general works |
recordtype |
marc |
publishDateSort |
2022 |
contenttype_str_mv |
txt |
container_start_page |
20139 |
author_browse |
Liu, Hongzhe Yan, Hua |
container_volume |
82 |
class |
070 004 VZ |
format_se |
Aufsätze |
author-letter |
Liu, Hongzhe |
doi_str_mv |
10.1007/s11042-022-14314-9 |
normlink |
(ORCID)0000-0001-9231-3175 |
normlink_prefix_str_mv |
(orcid)0000-0001-9231-3175 |
dewey-full |
070 004 |
title_sort |
an end-to-end multi-scale network based on autoencoder for infrared and visible image fusion |
title_auth |
An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion |
abstract |
Abstract Infrared and visible image fusion aims to obtain a more informative fusion image by merging the infrared and visible images. However, the existing methods have some shortcomings, such as detail information loss, unclear boundaries, and not being end-to-end. In this paper, we propose an end-to-end network architecture for infrared and visible image fusion task. Our network contains three essential parts: encoders, residual fusion module, and decoder. First, we input infrared and visible images to two encoders to extract shallow features, respectively. Subsequently, the two sets of features are concatenated and fed to the residual fusion module to extract multi-scale features and fuse them adequately. Finally, the fused image is obtained by the decoder. We conduct objective and subjective experiments on two public datasets. The comparison results with the state-of-art methods prove that the fusion results of the proposed method have better objective metrics and contain more detail information and more explicit boundary. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. |
abstractGer |
Abstract Infrared and visible image fusion aims to obtain a more informative fusion image by merging the infrared and visible images. However, the existing methods have some shortcomings, such as detail information loss, unclear boundaries, and not being end-to-end. In this paper, we propose an end-to-end network architecture for infrared and visible image fusion task. Our network contains three essential parts: encoders, residual fusion module, and decoder. First, we input infrared and visible images to two encoders to extract shallow features, respectively. Subsequently, the two sets of features are concatenated and fed to the residual fusion module to extract multi-scale features and fuse them adequately. Finally, the fused image is obtained by the decoder. We conduct objective and subjective experiments on two public datasets. The comparison results with the state-of-art methods prove that the fusion results of the proposed method have better objective metrics and contain more detail information and more explicit boundary. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. |
abstract_unstemmed |
Abstract Infrared and visible image fusion aims to obtain a more informative fusion image by merging the infrared and visible images. However, the existing methods have some shortcomings, such as detail information loss, unclear boundaries, and not being end-to-end. In this paper, we propose an end-to-end network architecture for infrared and visible image fusion task. Our network contains three essential parts: encoders, residual fusion module, and decoder. First, we input infrared and visible images to two encoders to extract shallow features, respectively. Subsequently, the two sets of features are concatenated and fed to the residual fusion module to extract multi-scale features and fuse them adequately. Finally, the fused image is obtained by the decoder. We conduct objective and subjective experiments on two public datasets. The comparison results with the state-of-art methods prove that the fusion results of the proposed method have better objective metrics and contain more detail information and more explicit boundary. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. |
collection_details |
GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT SSG-OLC-BUB SSG-OLC-MKW |
container_issue |
13 |
title_short |
An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion |
url |
https://doi.org/10.1007/s11042-022-14314-9 |
remote_bool |
false |
author2 |
Yan, Hua |
author2Str |
Yan, Hua |
ppnlink |
189064145 |
mediatype_str_mv |
n |
isOA_txt |
false |
hochschulschrift_bool |
false |
doi_str |
10.1007/s11042-022-14314-9 |
up_date |
2024-07-04T01:56:02.404Z |
_version_ |
1803611707707228160 |
fullrecord_marcxml |
<?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01000naa a22002652 4500</leader><controlfield tag="001">OLC2134624426</controlfield><controlfield tag="003">DE-627</controlfield><controlfield tag="005">20230510162757.0</controlfield><controlfield tag="007">tu</controlfield><controlfield tag="008">230510s2022 xx ||||| 00| ||eng c</controlfield><datafield tag="024" ind1="7" ind2=" "><subfield code="a">10.1007/s11042-022-14314-9</subfield><subfield code="2">doi</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-627)OLC2134624426</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-He213)s11042-022-14314-9-p</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-627</subfield><subfield code="b">ger</subfield><subfield code="c">DE-627</subfield><subfield code="e">rakwb</subfield></datafield><datafield tag="041" ind1=" " ind2=" "><subfield code="a">eng</subfield></datafield><datafield tag="082" ind1="0" ind2="4"><subfield code="a">070</subfield><subfield code="a">004</subfield><subfield code="q">VZ</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Liu, Hongzhe</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="c">2022</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">Text</subfield><subfield code="b">txt</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">ohne Hilfsmittel zu benutzen</subfield><subfield code="b">n</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">Band</subfield><subfield code="b">nc</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a">© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">Abstract Infrared and visible image fusion aims to obtain a more informative fusion image by merging the infrared and visible images. However, the existing methods have some shortcomings, such as detail information loss, unclear boundaries, and not being end-to-end. In this paper, we propose an end-to-end network architecture for infrared and visible image fusion task. Our network contains three essential parts: encoders, residual fusion module, and decoder. First, we input infrared and visible images to two encoders to extract shallow features, respectively. Subsequently, the two sets of features are concatenated and fed to the residual fusion module to extract multi-scale features and fuse them adequately. Finally, the fused image is obtained by the decoder. We conduct objective and subjective experiments on two public datasets. The comparison results with the state-of-art methods prove that the fusion results of the proposed method have better objective metrics and contain more detail information and more explicit boundary.</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Autoencoder</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Image fusion</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Infrared image</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Visible image</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Yan, Hua</subfield><subfield code="0">(orcid)0000-0001-9231-3175</subfield><subfield code="4">aut</subfield></datafield><datafield tag="773" ind1="0" ind2="8"><subfield code="i">Enthalten in</subfield><subfield code="t">Multimedia tools and applications</subfield><subfield code="d">Springer US, 1995</subfield><subfield code="g">82(2022), 13 vom: 27. Dez., Seite 20139-20156</subfield><subfield code="w">(DE-627)189064145</subfield><subfield code="w">(DE-600)1287642-2</subfield><subfield code="w">(DE-576)052842126</subfield><subfield code="x">1380-7501</subfield><subfield code="7">nnns</subfield></datafield><datafield tag="773" ind1="1" ind2="8"><subfield code="g">volume:82</subfield><subfield code="g">year:2022</subfield><subfield code="g">number:13</subfield><subfield code="g">day:27</subfield><subfield code="g">month:12</subfield><subfield code="g">pages:20139-20156</subfield></datafield><datafield tag="856" ind1="4" ind2="1"><subfield code="u">https://doi.org/10.1007/s11042-022-14314-9</subfield><subfield code="z">lizenzpflichtig</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_USEFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SYSFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_OLC</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-MAT</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-BUB</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-MKW</subfield></datafield><datafield tag="951" ind1=" " ind2=" "><subfield code="a">AR</subfield></datafield><datafield tag="952" ind1=" " ind2=" "><subfield code="d">82</subfield><subfield code="j">2022</subfield><subfield code="e">13</subfield><subfield code="b">27</subfield><subfield code="c">12</subfield><subfield code="h">20139-20156</subfield></datafield></record></collection>
|
score |
7.4009476 |