Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model
Previous studies have already shown that Raman spectroscopy can be used in the encoding of suspension array technology. However, almost all existing convolutional neural network-based decoding approaches rely on supervision with ground truth, and may not be well generalized to unseen datasets, which...
Ausführliche Beschreibung
Autor*in: |
Yu Yao [verfasserIn] Kaiwen Xue [verfasserIn] Liwang Liu [verfasserIn] Shanshan Zhu [verfasserIn] Chengfeng Yue [verfasserIn] Yanhong Ji [verfasserIn] |
---|
Format: |
E-Artikel |
---|---|
Sprache: |
Englisch |
Erschienen: |
2022 |
---|
Schlagwörter: |
---|
Übergeordnetes Werk: |
In: Journal of Innovative Optical Health Sciences - World Scientific Publishing, 2017, 15(2022), 04 |
---|---|
Übergeordnetes Werk: |
volume:15 ; year:2022 ; number:04 |
Links: |
Link aufrufen |
---|
DOI / URN: |
10.1142/S1793545822500250 |
---|
Katalog-ID: |
DOAJ079168078 |
---|
LEADER | 01000caa a22002652 4500 | ||
---|---|---|---|
001 | DOAJ079168078 | ||
003 | DE-627 | ||
005 | 20230503064335.0 | ||
007 | cr uuu---uuuuu | ||
008 | 230307s2022 xx |||||o 00| ||eng c | ||
024 | 7 | |a 10.1142/S1793545822500250 |2 doi | |
035 | |a (DE-627)DOAJ079168078 | ||
035 | |a (DE-599)DOAJbb2234f8ac5f4d4592c433baed65fc5c | ||
040 | |a DE-627 |b ger |c DE-627 |e rakwb | ||
041 | |a eng | ||
050 | 0 | |a QC350-467 | |
100 | 0 | |a Yu Yao |e verfasserin |4 aut | |
245 | 1 | 0 | |a Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model |
264 | 1 | |c 2022 | |
336 | |a Text |b txt |2 rdacontent | ||
337 | |a Computermedien |b c |2 rdamedia | ||
338 | |a Online-Ressource |b cr |2 rdacarrier | ||
520 | |a Previous studies have already shown that Raman spectroscopy can be used in the encoding of suspension array technology. However, almost all existing convolutional neural network-based decoding approaches rely on supervision with ground truth, and may not be well generalized to unseen datasets, which were collected under different experimental conditions, applying with the same coded material. In this study, we propose an improved model based on CyCADA, named as Detail constraint Cycle Domain Adaptive Model (DCDA). DCDA implements the classification of unseen datasets through domain adaptation, adapts representations at the encode level with decoder-share, and enforces coding features while leveraging a feat loss. To improve detailed structural constraints, DCDA takes downsample connection and skips connection. Our model improves the poor generalization of existing models and saves the cost of the labeling process for unseen target datasets. Compared with other models, extensive experiments and ablation studies show the superiority of DCDA in terms of classification stability and generalization. The model proposed by the research achieves a classification with an accuracy of 100% when applied in datasets, in which the spectrum in the source domain is far less than the target domain. | ||
650 | 4 | |a Domain adaption | |
650 | 4 | |a suspension arrays | |
650 | 4 | |a deep learning | |
650 | 4 | |a Raman spectrum | |
650 | 4 | |a generalization | |
653 | 0 | |a Technology | |
653 | 0 | |a T | |
653 | 0 | |a Optics. Light | |
700 | 0 | |a Kaiwen Xue |e verfasserin |4 aut | |
700 | 0 | |a Liwang Liu |e verfasserin |4 aut | |
700 | 0 | |a Shanshan Zhu |e verfasserin |4 aut | |
700 | 0 | |a Chengfeng Yue |e verfasserin |4 aut | |
700 | 0 | |a Yanhong Ji |e verfasserin |4 aut | |
773 | 0 | 8 | |i In |t Journal of Innovative Optical Health Sciences |d World Scientific Publishing, 2017 |g 15(2022), 04 |w (DE-627)60940315X |w (DE-600)2515441-2 |x 17937205 |7 nnns |
773 | 1 | 8 | |g volume:15 |g year:2022 |g number:04 |
856 | 4 | 0 | |u https://doi.org/10.1142/S1793545822500250 |z kostenfrei |
856 | 4 | 0 | |u https://doaj.org/article/bb2234f8ac5f4d4592c433baed65fc5c |z kostenfrei |
856 | 4 | 0 | |u https://www.worldscientific.com/doi/10.1142/S1793545822500250 |z kostenfrei |
856 | 4 | 2 | |u https://doaj.org/toc/1793-5458 |y Journal toc |z kostenfrei |
856 | 4 | 2 | |u https://doaj.org/toc/1793-7205 |y Journal toc |z kostenfrei |
912 | |a GBV_USEFLAG_A | ||
912 | |a SYSFLAG_A | ||
912 | |a GBV_DOAJ | ||
912 | |a SSG-OLC-PHA | ||
912 | |a GBV_ILN_11 | ||
912 | |a GBV_ILN_20 | ||
912 | |a GBV_ILN_22 | ||
912 | |a GBV_ILN_23 | ||
912 | |a GBV_ILN_24 | ||
912 | |a GBV_ILN_31 | ||
912 | |a GBV_ILN_39 | ||
912 | |a GBV_ILN_40 | ||
912 | |a GBV_ILN_60 | ||
912 | |a GBV_ILN_62 | ||
912 | |a GBV_ILN_63 | ||
912 | |a GBV_ILN_65 | ||
912 | |a GBV_ILN_69 | ||
912 | |a GBV_ILN_70 | ||
912 | |a GBV_ILN_73 | ||
912 | |a GBV_ILN_74 | ||
912 | |a GBV_ILN_95 | ||
912 | |a GBV_ILN_105 | ||
912 | |a GBV_ILN_110 | ||
912 | |a GBV_ILN_151 | ||
912 | |a GBV_ILN_161 | ||
912 | |a GBV_ILN_170 | ||
912 | |a GBV_ILN_206 | ||
912 | |a GBV_ILN_213 | ||
912 | |a GBV_ILN_230 | ||
912 | |a GBV_ILN_285 | ||
912 | |a GBV_ILN_293 | ||
912 | |a GBV_ILN_370 | ||
912 | |a GBV_ILN_602 | ||
912 | |a GBV_ILN_2005 | ||
912 | |a GBV_ILN_2009 | ||
912 | |a GBV_ILN_2011 | ||
912 | |a GBV_ILN_2014 | ||
912 | |a GBV_ILN_2055 | ||
912 | |a GBV_ILN_2111 | ||
912 | |a GBV_ILN_4012 | ||
912 | |a GBV_ILN_4037 | ||
912 | |a GBV_ILN_4112 | ||
912 | |a GBV_ILN_4125 | ||
912 | |a GBV_ILN_4126 | ||
912 | |a GBV_ILN_4249 | ||
912 | |a GBV_ILN_4305 | ||
912 | |a GBV_ILN_4306 | ||
912 | |a GBV_ILN_4307 | ||
912 | |a GBV_ILN_4313 | ||
912 | |a GBV_ILN_4322 | ||
912 | |a GBV_ILN_4323 | ||
912 | |a GBV_ILN_4324 | ||
912 | |a GBV_ILN_4325 | ||
912 | |a GBV_ILN_4335 | ||
912 | |a GBV_ILN_4338 | ||
912 | |a GBV_ILN_4367 | ||
912 | |a GBV_ILN_4700 | ||
951 | |a AR | ||
952 | |d 15 |j 2022 |e 04 |
author_variant |
y y yy k x kx l l ll s z sz c y cy y j yj |
---|---|
matchkey_str |
article:17937205:2022----::eoigfaasetocpecddupninrasaeoteealosri |
hierarchy_sort_str |
2022 |
callnumber-subject-code |
QC |
publishDate |
2022 |
allfields |
10.1142/S1793545822500250 doi (DE-627)DOAJ079168078 (DE-599)DOAJbb2234f8ac5f4d4592c433baed65fc5c DE-627 ger DE-627 rakwb eng QC350-467 Yu Yao verfasserin aut Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model 2022 Text txt rdacontent Computermedien c rdamedia Online-Ressource cr rdacarrier Previous studies have already shown that Raman spectroscopy can be used in the encoding of suspension array technology. However, almost all existing convolutional neural network-based decoding approaches rely on supervision with ground truth, and may not be well generalized to unseen datasets, which were collected under different experimental conditions, applying with the same coded material. In this study, we propose an improved model based on CyCADA, named as Detail constraint Cycle Domain Adaptive Model (DCDA). DCDA implements the classification of unseen datasets through domain adaptation, adapts representations at the encode level with decoder-share, and enforces coding features while leveraging a feat loss. To improve detailed structural constraints, DCDA takes downsample connection and skips connection. Our model improves the poor generalization of existing models and saves the cost of the labeling process for unseen target datasets. Compared with other models, extensive experiments and ablation studies show the superiority of DCDA in terms of classification stability and generalization. The model proposed by the research achieves a classification with an accuracy of 100% when applied in datasets, in which the spectrum in the source domain is far less than the target domain. Domain adaption suspension arrays deep learning Raman spectrum generalization Technology T Optics. Light Kaiwen Xue verfasserin aut Liwang Liu verfasserin aut Shanshan Zhu verfasserin aut Chengfeng Yue verfasserin aut Yanhong Ji verfasserin aut In Journal of Innovative Optical Health Sciences World Scientific Publishing, 2017 15(2022), 04 (DE-627)60940315X (DE-600)2515441-2 17937205 nnns volume:15 year:2022 number:04 https://doi.org/10.1142/S1793545822500250 kostenfrei https://doaj.org/article/bb2234f8ac5f4d4592c433baed65fc5c kostenfrei https://www.worldscientific.com/doi/10.1142/S1793545822500250 kostenfrei https://doaj.org/toc/1793-5458 Journal toc kostenfrei https://doaj.org/toc/1793-7205 Journal toc kostenfrei GBV_USEFLAG_A SYSFLAG_A GBV_DOAJ SSG-OLC-PHA GBV_ILN_11 GBV_ILN_20 GBV_ILN_22 GBV_ILN_23 GBV_ILN_24 GBV_ILN_31 GBV_ILN_39 GBV_ILN_40 GBV_ILN_60 GBV_ILN_62 GBV_ILN_63 GBV_ILN_65 GBV_ILN_69 GBV_ILN_70 GBV_ILN_73 GBV_ILN_74 GBV_ILN_95 GBV_ILN_105 GBV_ILN_110 GBV_ILN_151 GBV_ILN_161 GBV_ILN_170 GBV_ILN_206 GBV_ILN_213 GBV_ILN_230 GBV_ILN_285 GBV_ILN_293 GBV_ILN_370 GBV_ILN_602 GBV_ILN_2005 GBV_ILN_2009 GBV_ILN_2011 GBV_ILN_2014 GBV_ILN_2055 GBV_ILN_2111 GBV_ILN_4012 GBV_ILN_4037 GBV_ILN_4112 GBV_ILN_4125 GBV_ILN_4126 GBV_ILN_4249 GBV_ILN_4305 GBV_ILN_4306 GBV_ILN_4307 GBV_ILN_4313 GBV_ILN_4322 GBV_ILN_4323 GBV_ILN_4324 GBV_ILN_4325 GBV_ILN_4335 GBV_ILN_4338 GBV_ILN_4367 GBV_ILN_4700 AR 15 2022 04 |
spelling |
10.1142/S1793545822500250 doi (DE-627)DOAJ079168078 (DE-599)DOAJbb2234f8ac5f4d4592c433baed65fc5c DE-627 ger DE-627 rakwb eng QC350-467 Yu Yao verfasserin aut Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model 2022 Text txt rdacontent Computermedien c rdamedia Online-Ressource cr rdacarrier Previous studies have already shown that Raman spectroscopy can be used in the encoding of suspension array technology. However, almost all existing convolutional neural network-based decoding approaches rely on supervision with ground truth, and may not be well generalized to unseen datasets, which were collected under different experimental conditions, applying with the same coded material. In this study, we propose an improved model based on CyCADA, named as Detail constraint Cycle Domain Adaptive Model (DCDA). DCDA implements the classification of unseen datasets through domain adaptation, adapts representations at the encode level with decoder-share, and enforces coding features while leveraging a feat loss. To improve detailed structural constraints, DCDA takes downsample connection and skips connection. Our model improves the poor generalization of existing models and saves the cost of the labeling process for unseen target datasets. Compared with other models, extensive experiments and ablation studies show the superiority of DCDA in terms of classification stability and generalization. The model proposed by the research achieves a classification with an accuracy of 100% when applied in datasets, in which the spectrum in the source domain is far less than the target domain. Domain adaption suspension arrays deep learning Raman spectrum generalization Technology T Optics. Light Kaiwen Xue verfasserin aut Liwang Liu verfasserin aut Shanshan Zhu verfasserin aut Chengfeng Yue verfasserin aut Yanhong Ji verfasserin aut In Journal of Innovative Optical Health Sciences World Scientific Publishing, 2017 15(2022), 04 (DE-627)60940315X (DE-600)2515441-2 17937205 nnns volume:15 year:2022 number:04 https://doi.org/10.1142/S1793545822500250 kostenfrei https://doaj.org/article/bb2234f8ac5f4d4592c433baed65fc5c kostenfrei https://www.worldscientific.com/doi/10.1142/S1793545822500250 kostenfrei https://doaj.org/toc/1793-5458 Journal toc kostenfrei https://doaj.org/toc/1793-7205 Journal toc kostenfrei GBV_USEFLAG_A SYSFLAG_A GBV_DOAJ SSG-OLC-PHA GBV_ILN_11 GBV_ILN_20 GBV_ILN_22 GBV_ILN_23 GBV_ILN_24 GBV_ILN_31 GBV_ILN_39 GBV_ILN_40 GBV_ILN_60 GBV_ILN_62 GBV_ILN_63 GBV_ILN_65 GBV_ILN_69 GBV_ILN_70 GBV_ILN_73 GBV_ILN_74 GBV_ILN_95 GBV_ILN_105 GBV_ILN_110 GBV_ILN_151 GBV_ILN_161 GBV_ILN_170 GBV_ILN_206 GBV_ILN_213 GBV_ILN_230 GBV_ILN_285 GBV_ILN_293 GBV_ILN_370 GBV_ILN_602 GBV_ILN_2005 GBV_ILN_2009 GBV_ILN_2011 GBV_ILN_2014 GBV_ILN_2055 GBV_ILN_2111 GBV_ILN_4012 GBV_ILN_4037 GBV_ILN_4112 GBV_ILN_4125 GBV_ILN_4126 GBV_ILN_4249 GBV_ILN_4305 GBV_ILN_4306 GBV_ILN_4307 GBV_ILN_4313 GBV_ILN_4322 GBV_ILN_4323 GBV_ILN_4324 GBV_ILN_4325 GBV_ILN_4335 GBV_ILN_4338 GBV_ILN_4367 GBV_ILN_4700 AR 15 2022 04 |
allfields_unstemmed |
10.1142/S1793545822500250 doi (DE-627)DOAJ079168078 (DE-599)DOAJbb2234f8ac5f4d4592c433baed65fc5c DE-627 ger DE-627 rakwb eng QC350-467 Yu Yao verfasserin aut Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model 2022 Text txt rdacontent Computermedien c rdamedia Online-Ressource cr rdacarrier Previous studies have already shown that Raman spectroscopy can be used in the encoding of suspension array technology. However, almost all existing convolutional neural network-based decoding approaches rely on supervision with ground truth, and may not be well generalized to unseen datasets, which were collected under different experimental conditions, applying with the same coded material. In this study, we propose an improved model based on CyCADA, named as Detail constraint Cycle Domain Adaptive Model (DCDA). DCDA implements the classification of unseen datasets through domain adaptation, adapts representations at the encode level with decoder-share, and enforces coding features while leveraging a feat loss. To improve detailed structural constraints, DCDA takes downsample connection and skips connection. Our model improves the poor generalization of existing models and saves the cost of the labeling process for unseen target datasets. Compared with other models, extensive experiments and ablation studies show the superiority of DCDA in terms of classification stability and generalization. The model proposed by the research achieves a classification with an accuracy of 100% when applied in datasets, in which the spectrum in the source domain is far less than the target domain. Domain adaption suspension arrays deep learning Raman spectrum generalization Technology T Optics. Light Kaiwen Xue verfasserin aut Liwang Liu verfasserin aut Shanshan Zhu verfasserin aut Chengfeng Yue verfasserin aut Yanhong Ji verfasserin aut In Journal of Innovative Optical Health Sciences World Scientific Publishing, 2017 15(2022), 04 (DE-627)60940315X (DE-600)2515441-2 17937205 nnns volume:15 year:2022 number:04 https://doi.org/10.1142/S1793545822500250 kostenfrei https://doaj.org/article/bb2234f8ac5f4d4592c433baed65fc5c kostenfrei https://www.worldscientific.com/doi/10.1142/S1793545822500250 kostenfrei https://doaj.org/toc/1793-5458 Journal toc kostenfrei https://doaj.org/toc/1793-7205 Journal toc kostenfrei GBV_USEFLAG_A SYSFLAG_A GBV_DOAJ SSG-OLC-PHA GBV_ILN_11 GBV_ILN_20 GBV_ILN_22 GBV_ILN_23 GBV_ILN_24 GBV_ILN_31 GBV_ILN_39 GBV_ILN_40 GBV_ILN_60 GBV_ILN_62 GBV_ILN_63 GBV_ILN_65 GBV_ILN_69 GBV_ILN_70 GBV_ILN_73 GBV_ILN_74 GBV_ILN_95 GBV_ILN_105 GBV_ILN_110 GBV_ILN_151 GBV_ILN_161 GBV_ILN_170 GBV_ILN_206 GBV_ILN_213 GBV_ILN_230 GBV_ILN_285 GBV_ILN_293 GBV_ILN_370 GBV_ILN_602 GBV_ILN_2005 GBV_ILN_2009 GBV_ILN_2011 GBV_ILN_2014 GBV_ILN_2055 GBV_ILN_2111 GBV_ILN_4012 GBV_ILN_4037 GBV_ILN_4112 GBV_ILN_4125 GBV_ILN_4126 GBV_ILN_4249 GBV_ILN_4305 GBV_ILN_4306 GBV_ILN_4307 GBV_ILN_4313 GBV_ILN_4322 GBV_ILN_4323 GBV_ILN_4324 GBV_ILN_4325 GBV_ILN_4335 GBV_ILN_4338 GBV_ILN_4367 GBV_ILN_4700 AR 15 2022 04 |
allfieldsGer |
10.1142/S1793545822500250 doi (DE-627)DOAJ079168078 (DE-599)DOAJbb2234f8ac5f4d4592c433baed65fc5c DE-627 ger DE-627 rakwb eng QC350-467 Yu Yao verfasserin aut Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model 2022 Text txt rdacontent Computermedien c rdamedia Online-Ressource cr rdacarrier Previous studies have already shown that Raman spectroscopy can be used in the encoding of suspension array technology. However, almost all existing convolutional neural network-based decoding approaches rely on supervision with ground truth, and may not be well generalized to unseen datasets, which were collected under different experimental conditions, applying with the same coded material. In this study, we propose an improved model based on CyCADA, named as Detail constraint Cycle Domain Adaptive Model (DCDA). DCDA implements the classification of unseen datasets through domain adaptation, adapts representations at the encode level with decoder-share, and enforces coding features while leveraging a feat loss. To improve detailed structural constraints, DCDA takes downsample connection and skips connection. Our model improves the poor generalization of existing models and saves the cost of the labeling process for unseen target datasets. Compared with other models, extensive experiments and ablation studies show the superiority of DCDA in terms of classification stability and generalization. The model proposed by the research achieves a classification with an accuracy of 100% when applied in datasets, in which the spectrum in the source domain is far less than the target domain. Domain adaption suspension arrays deep learning Raman spectrum generalization Technology T Optics. Light Kaiwen Xue verfasserin aut Liwang Liu verfasserin aut Shanshan Zhu verfasserin aut Chengfeng Yue verfasserin aut Yanhong Ji verfasserin aut In Journal of Innovative Optical Health Sciences World Scientific Publishing, 2017 15(2022), 04 (DE-627)60940315X (DE-600)2515441-2 17937205 nnns volume:15 year:2022 number:04 https://doi.org/10.1142/S1793545822500250 kostenfrei https://doaj.org/article/bb2234f8ac5f4d4592c433baed65fc5c kostenfrei https://www.worldscientific.com/doi/10.1142/S1793545822500250 kostenfrei https://doaj.org/toc/1793-5458 Journal toc kostenfrei https://doaj.org/toc/1793-7205 Journal toc kostenfrei GBV_USEFLAG_A SYSFLAG_A GBV_DOAJ SSG-OLC-PHA GBV_ILN_11 GBV_ILN_20 GBV_ILN_22 GBV_ILN_23 GBV_ILN_24 GBV_ILN_31 GBV_ILN_39 GBV_ILN_40 GBV_ILN_60 GBV_ILN_62 GBV_ILN_63 GBV_ILN_65 GBV_ILN_69 GBV_ILN_70 GBV_ILN_73 GBV_ILN_74 GBV_ILN_95 GBV_ILN_105 GBV_ILN_110 GBV_ILN_151 GBV_ILN_161 GBV_ILN_170 GBV_ILN_206 GBV_ILN_213 GBV_ILN_230 GBV_ILN_285 GBV_ILN_293 GBV_ILN_370 GBV_ILN_602 GBV_ILN_2005 GBV_ILN_2009 GBV_ILN_2011 GBV_ILN_2014 GBV_ILN_2055 GBV_ILN_2111 GBV_ILN_4012 GBV_ILN_4037 GBV_ILN_4112 GBV_ILN_4125 GBV_ILN_4126 GBV_ILN_4249 GBV_ILN_4305 GBV_ILN_4306 GBV_ILN_4307 GBV_ILN_4313 GBV_ILN_4322 GBV_ILN_4323 GBV_ILN_4324 GBV_ILN_4325 GBV_ILN_4335 GBV_ILN_4338 GBV_ILN_4367 GBV_ILN_4700 AR 15 2022 04 |
allfieldsSound |
10.1142/S1793545822500250 doi (DE-627)DOAJ079168078 (DE-599)DOAJbb2234f8ac5f4d4592c433baed65fc5c DE-627 ger DE-627 rakwb eng QC350-467 Yu Yao verfasserin aut Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model 2022 Text txt rdacontent Computermedien c rdamedia Online-Ressource cr rdacarrier Previous studies have already shown that Raman spectroscopy can be used in the encoding of suspension array technology. However, almost all existing convolutional neural network-based decoding approaches rely on supervision with ground truth, and may not be well generalized to unseen datasets, which were collected under different experimental conditions, applying with the same coded material. In this study, we propose an improved model based on CyCADA, named as Detail constraint Cycle Domain Adaptive Model (DCDA). DCDA implements the classification of unseen datasets through domain adaptation, adapts representations at the encode level with decoder-share, and enforces coding features while leveraging a feat loss. To improve detailed structural constraints, DCDA takes downsample connection and skips connection. Our model improves the poor generalization of existing models and saves the cost of the labeling process for unseen target datasets. Compared with other models, extensive experiments and ablation studies show the superiority of DCDA in terms of classification stability and generalization. The model proposed by the research achieves a classification with an accuracy of 100% when applied in datasets, in which the spectrum in the source domain is far less than the target domain. Domain adaption suspension arrays deep learning Raman spectrum generalization Technology T Optics. Light Kaiwen Xue verfasserin aut Liwang Liu verfasserin aut Shanshan Zhu verfasserin aut Chengfeng Yue verfasserin aut Yanhong Ji verfasserin aut In Journal of Innovative Optical Health Sciences World Scientific Publishing, 2017 15(2022), 04 (DE-627)60940315X (DE-600)2515441-2 17937205 nnns volume:15 year:2022 number:04 https://doi.org/10.1142/S1793545822500250 kostenfrei https://doaj.org/article/bb2234f8ac5f4d4592c433baed65fc5c kostenfrei https://www.worldscientific.com/doi/10.1142/S1793545822500250 kostenfrei https://doaj.org/toc/1793-5458 Journal toc kostenfrei https://doaj.org/toc/1793-7205 Journal toc kostenfrei GBV_USEFLAG_A SYSFLAG_A GBV_DOAJ SSG-OLC-PHA GBV_ILN_11 GBV_ILN_20 GBV_ILN_22 GBV_ILN_23 GBV_ILN_24 GBV_ILN_31 GBV_ILN_39 GBV_ILN_40 GBV_ILN_60 GBV_ILN_62 GBV_ILN_63 GBV_ILN_65 GBV_ILN_69 GBV_ILN_70 GBV_ILN_73 GBV_ILN_74 GBV_ILN_95 GBV_ILN_105 GBV_ILN_110 GBV_ILN_151 GBV_ILN_161 GBV_ILN_170 GBV_ILN_206 GBV_ILN_213 GBV_ILN_230 GBV_ILN_285 GBV_ILN_293 GBV_ILN_370 GBV_ILN_602 GBV_ILN_2005 GBV_ILN_2009 GBV_ILN_2011 GBV_ILN_2014 GBV_ILN_2055 GBV_ILN_2111 GBV_ILN_4012 GBV_ILN_4037 GBV_ILN_4112 GBV_ILN_4125 GBV_ILN_4126 GBV_ILN_4249 GBV_ILN_4305 GBV_ILN_4306 GBV_ILN_4307 GBV_ILN_4313 GBV_ILN_4322 GBV_ILN_4323 GBV_ILN_4324 GBV_ILN_4325 GBV_ILN_4335 GBV_ILN_4338 GBV_ILN_4367 GBV_ILN_4700 AR 15 2022 04 |
language |
English |
source |
In Journal of Innovative Optical Health Sciences 15(2022), 04 volume:15 year:2022 number:04 |
sourceStr |
In Journal of Innovative Optical Health Sciences 15(2022), 04 volume:15 year:2022 number:04 |
format_phy_str_mv |
Article |
institution |
findex.gbv.de |
topic_facet |
Domain adaption suspension arrays deep learning Raman spectrum generalization Technology T Optics. Light |
isfreeaccess_bool |
true |
container_title |
Journal of Innovative Optical Health Sciences |
authorswithroles_txt_mv |
Yu Yao @@aut@@ Kaiwen Xue @@aut@@ Liwang Liu @@aut@@ Shanshan Zhu @@aut@@ Chengfeng Yue @@aut@@ Yanhong Ji @@aut@@ |
publishDateDaySort_date |
2022-01-01T00:00:00Z |
hierarchy_top_id |
60940315X |
id |
DOAJ079168078 |
language_de |
englisch |
fullrecord |
<?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01000caa a22002652 4500</leader><controlfield tag="001">DOAJ079168078</controlfield><controlfield tag="003">DE-627</controlfield><controlfield tag="005">20230503064335.0</controlfield><controlfield tag="007">cr uuu---uuuuu</controlfield><controlfield tag="008">230307s2022 xx |||||o 00| ||eng c</controlfield><datafield tag="024" ind1="7" ind2=" "><subfield code="a">10.1142/S1793545822500250</subfield><subfield code="2">doi</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-627)DOAJ079168078</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-599)DOAJbb2234f8ac5f4d4592c433baed65fc5c</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-627</subfield><subfield code="b">ger</subfield><subfield code="c">DE-627</subfield><subfield code="e">rakwb</subfield></datafield><datafield tag="041" ind1=" " ind2=" "><subfield code="a">eng</subfield></datafield><datafield tag="050" ind1=" " ind2="0"><subfield code="a">QC350-467</subfield></datafield><datafield tag="100" ind1="0" ind2=" "><subfield code="a">Yu Yao</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="c">2022</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">Text</subfield><subfield code="b">txt</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">Computermedien</subfield><subfield code="b">c</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">Online-Ressource</subfield><subfield code="b">cr</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">Previous studies have already shown that Raman spectroscopy can be used in the encoding of suspension array technology. However, almost all existing convolutional neural network-based decoding approaches rely on supervision with ground truth, and may not be well generalized to unseen datasets, which were collected under different experimental conditions, applying with the same coded material. In this study, we propose an improved model based on CyCADA, named as Detail constraint Cycle Domain Adaptive Model (DCDA). DCDA implements the classification of unseen datasets through domain adaptation, adapts representations at the encode level with decoder-share, and enforces coding features while leveraging a feat loss. To improve detailed structural constraints, DCDA takes downsample connection and skips connection. Our model improves the poor generalization of existing models and saves the cost of the labeling process for unseen target datasets. Compared with other models, extensive experiments and ablation studies show the superiority of DCDA in terms of classification stability and generalization. The model proposed by the research achieves a classification with an accuracy of 100% when applied in datasets, in which the spectrum in the source domain is far less than the target domain.</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Domain adaption</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">suspension arrays</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">deep learning</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Raman spectrum</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">generalization</subfield></datafield><datafield tag="653" ind1=" " ind2="0"><subfield code="a">Technology</subfield></datafield><datafield tag="653" ind1=" " ind2="0"><subfield code="a">T</subfield></datafield><datafield tag="653" ind1=" " ind2="0"><subfield code="a">Optics. Light</subfield></datafield><datafield tag="700" ind1="0" ind2=" "><subfield code="a">Kaiwen Xue</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="0" ind2=" "><subfield code="a">Liwang Liu</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="0" ind2=" "><subfield code="a">Shanshan Zhu</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="0" ind2=" "><subfield code="a">Chengfeng Yue</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="0" ind2=" "><subfield code="a">Yanhong Ji</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="773" ind1="0" ind2="8"><subfield code="i">In</subfield><subfield code="t">Journal of Innovative Optical Health Sciences</subfield><subfield code="d">World Scientific Publishing, 2017</subfield><subfield code="g">15(2022), 04</subfield><subfield code="w">(DE-627)60940315X</subfield><subfield code="w">(DE-600)2515441-2</subfield><subfield code="x">17937205</subfield><subfield code="7">nnns</subfield></datafield><datafield tag="773" ind1="1" ind2="8"><subfield code="g">volume:15</subfield><subfield code="g">year:2022</subfield><subfield code="g">number:04</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="u">https://doi.org/10.1142/S1793545822500250</subfield><subfield code="z">kostenfrei</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="u">https://doaj.org/article/bb2234f8ac5f4d4592c433baed65fc5c</subfield><subfield code="z">kostenfrei</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="u">https://www.worldscientific.com/doi/10.1142/S1793545822500250</subfield><subfield code="z">kostenfrei</subfield></datafield><datafield tag="856" ind1="4" ind2="2"><subfield code="u">https://doaj.org/toc/1793-5458</subfield><subfield code="y">Journal toc</subfield><subfield code="z">kostenfrei</subfield></datafield><datafield tag="856" ind1="4" ind2="2"><subfield code="u">https://doaj.org/toc/1793-7205</subfield><subfield code="y">Journal toc</subfield><subfield code="z">kostenfrei</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_USEFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SYSFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_DOAJ</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-PHA</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_11</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_20</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_22</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_23</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_24</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_31</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_39</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_40</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_60</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_62</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_63</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_65</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_69</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_70</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_73</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_74</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_95</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_105</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_110</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_151</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_161</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_170</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_206</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_213</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_230</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_285</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_293</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_370</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_602</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_2005</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_2009</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_2011</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_2014</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_2055</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_2111</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4012</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4037</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4112</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4125</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4126</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4249</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4305</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4306</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4307</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4313</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4322</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4323</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4324</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4325</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4335</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4338</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4367</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4700</subfield></datafield><datafield tag="951" ind1=" " ind2=" "><subfield code="a">AR</subfield></datafield><datafield tag="952" ind1=" " ind2=" "><subfield code="d">15</subfield><subfield code="j">2022</subfield><subfield code="e">04</subfield></datafield></record></collection>
|
callnumber-first |
Q - Science |
author |
Yu Yao |
spellingShingle |
Yu Yao misc QC350-467 misc Domain adaption misc suspension arrays misc deep learning misc Raman spectrum misc generalization misc Technology misc T misc Optics. Light Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model |
authorStr |
Yu Yao |
ppnlink_with_tag_str_mv |
@@773@@(DE-627)60940315X |
format |
electronic Article |
delete_txt_mv |
keep |
author_role |
aut aut aut aut aut aut |
collection |
DOAJ |
remote_str |
true |
callnumber-label |
QC350-467 |
illustrated |
Not Illustrated |
issn |
17937205 |
topic_title |
QC350-467 Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model Domain adaption suspension arrays deep learning Raman spectrum generalization |
topic |
misc QC350-467 misc Domain adaption misc suspension arrays misc deep learning misc Raman spectrum misc generalization misc Technology misc T misc Optics. Light |
topic_unstemmed |
misc QC350-467 misc Domain adaption misc suspension arrays misc deep learning misc Raman spectrum misc generalization misc Technology misc T misc Optics. Light |
topic_browse |
misc QC350-467 misc Domain adaption misc suspension arrays misc deep learning misc Raman spectrum misc generalization misc Technology misc T misc Optics. Light |
format_facet |
Elektronische Aufsätze Aufsätze Elektronische Ressource |
format_main_str_mv |
Text Zeitschrift/Artikel |
carriertype_str_mv |
cr |
hierarchy_parent_title |
Journal of Innovative Optical Health Sciences |
hierarchy_parent_id |
60940315X |
hierarchy_top_title |
Journal of Innovative Optical Health Sciences |
isfreeaccess_txt |
true |
familylinks_str_mv |
(DE-627)60940315X (DE-600)2515441-2 |
title |
Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model |
ctrlnum |
(DE-627)DOAJ079168078 (DE-599)DOAJbb2234f8ac5f4d4592c433baed65fc5c |
title_full |
Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model |
author_sort |
Yu Yao |
journal |
Journal of Innovative Optical Health Sciences |
journalStr |
Journal of Innovative Optical Health Sciences |
callnumber-first-code |
Q |
lang_code |
eng |
isOA_bool |
true |
recordtype |
marc |
publishDateSort |
2022 |
contenttype_str_mv |
txt |
author_browse |
Yu Yao Kaiwen Xue Liwang Liu Shanshan Zhu Chengfeng Yue Yanhong Ji |
container_volume |
15 |
class |
QC350-467 |
format_se |
Elektronische Aufsätze |
author-letter |
Yu Yao |
doi_str_mv |
10.1142/S1793545822500250 |
author2-role |
verfasserin |
title_sort |
decoding of raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model |
callnumber |
QC350-467 |
title_auth |
Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model |
abstract |
Previous studies have already shown that Raman spectroscopy can be used in the encoding of suspension array technology. However, almost all existing convolutional neural network-based decoding approaches rely on supervision with ground truth, and may not be well generalized to unseen datasets, which were collected under different experimental conditions, applying with the same coded material. In this study, we propose an improved model based on CyCADA, named as Detail constraint Cycle Domain Adaptive Model (DCDA). DCDA implements the classification of unseen datasets through domain adaptation, adapts representations at the encode level with decoder-share, and enforces coding features while leveraging a feat loss. To improve detailed structural constraints, DCDA takes downsample connection and skips connection. Our model improves the poor generalization of existing models and saves the cost of the labeling process for unseen target datasets. Compared with other models, extensive experiments and ablation studies show the superiority of DCDA in terms of classification stability and generalization. The model proposed by the research achieves a classification with an accuracy of 100% when applied in datasets, in which the spectrum in the source domain is far less than the target domain. |
abstractGer |
Previous studies have already shown that Raman spectroscopy can be used in the encoding of suspension array technology. However, almost all existing convolutional neural network-based decoding approaches rely on supervision with ground truth, and may not be well generalized to unseen datasets, which were collected under different experimental conditions, applying with the same coded material. In this study, we propose an improved model based on CyCADA, named as Detail constraint Cycle Domain Adaptive Model (DCDA). DCDA implements the classification of unseen datasets through domain adaptation, adapts representations at the encode level with decoder-share, and enforces coding features while leveraging a feat loss. To improve detailed structural constraints, DCDA takes downsample connection and skips connection. Our model improves the poor generalization of existing models and saves the cost of the labeling process for unseen target datasets. Compared with other models, extensive experiments and ablation studies show the superiority of DCDA in terms of classification stability and generalization. The model proposed by the research achieves a classification with an accuracy of 100% when applied in datasets, in which the spectrum in the source domain is far less than the target domain. |
abstract_unstemmed |
Previous studies have already shown that Raman spectroscopy can be used in the encoding of suspension array technology. However, almost all existing convolutional neural network-based decoding approaches rely on supervision with ground truth, and may not be well generalized to unseen datasets, which were collected under different experimental conditions, applying with the same coded material. In this study, we propose an improved model based on CyCADA, named as Detail constraint Cycle Domain Adaptive Model (DCDA). DCDA implements the classification of unseen datasets through domain adaptation, adapts representations at the encode level with decoder-share, and enforces coding features while leveraging a feat loss. To improve detailed structural constraints, DCDA takes downsample connection and skips connection. Our model improves the poor generalization of existing models and saves the cost of the labeling process for unseen target datasets. Compared with other models, extensive experiments and ablation studies show the superiority of DCDA in terms of classification stability and generalization. The model proposed by the research achieves a classification with an accuracy of 100% when applied in datasets, in which the spectrum in the source domain is far less than the target domain. |
collection_details |
GBV_USEFLAG_A SYSFLAG_A GBV_DOAJ SSG-OLC-PHA GBV_ILN_11 GBV_ILN_20 GBV_ILN_22 GBV_ILN_23 GBV_ILN_24 GBV_ILN_31 GBV_ILN_39 GBV_ILN_40 GBV_ILN_60 GBV_ILN_62 GBV_ILN_63 GBV_ILN_65 GBV_ILN_69 GBV_ILN_70 GBV_ILN_73 GBV_ILN_74 GBV_ILN_95 GBV_ILN_105 GBV_ILN_110 GBV_ILN_151 GBV_ILN_161 GBV_ILN_170 GBV_ILN_206 GBV_ILN_213 GBV_ILN_230 GBV_ILN_285 GBV_ILN_293 GBV_ILN_370 GBV_ILN_602 GBV_ILN_2005 GBV_ILN_2009 GBV_ILN_2011 GBV_ILN_2014 GBV_ILN_2055 GBV_ILN_2111 GBV_ILN_4012 GBV_ILN_4037 GBV_ILN_4112 GBV_ILN_4125 GBV_ILN_4126 GBV_ILN_4249 GBV_ILN_4305 GBV_ILN_4306 GBV_ILN_4307 GBV_ILN_4313 GBV_ILN_4322 GBV_ILN_4323 GBV_ILN_4324 GBV_ILN_4325 GBV_ILN_4335 GBV_ILN_4338 GBV_ILN_4367 GBV_ILN_4700 |
container_issue |
04 |
title_short |
Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model |
url |
https://doi.org/10.1142/S1793545822500250 https://doaj.org/article/bb2234f8ac5f4d4592c433baed65fc5c https://www.worldscientific.com/doi/10.1142/S1793545822500250 https://doaj.org/toc/1793-5458 https://doaj.org/toc/1793-7205 |
remote_bool |
true |
author2 |
Kaiwen Xue Liwang Liu Shanshan Zhu Chengfeng Yue Yanhong Ji |
author2Str |
Kaiwen Xue Liwang Liu Shanshan Zhu Chengfeng Yue Yanhong Ji |
ppnlink |
60940315X |
callnumber-subject |
QC - Physics |
mediatype_str_mv |
c |
isOA_txt |
true |
hochschulschrift_bool |
false |
doi_str |
10.1142/S1793545822500250 |
callnumber-a |
QC350-467 |
up_date |
2024-07-03T22:04:18.376Z |
_version_ |
1803597128275066880 |
fullrecord_marcxml |
<?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01000caa a22002652 4500</leader><controlfield tag="001">DOAJ079168078</controlfield><controlfield tag="003">DE-627</controlfield><controlfield tag="005">20230503064335.0</controlfield><controlfield tag="007">cr uuu---uuuuu</controlfield><controlfield tag="008">230307s2022 xx |||||o 00| ||eng c</controlfield><datafield tag="024" ind1="7" ind2=" "><subfield code="a">10.1142/S1793545822500250</subfield><subfield code="2">doi</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-627)DOAJ079168078</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-599)DOAJbb2234f8ac5f4d4592c433baed65fc5c</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-627</subfield><subfield code="b">ger</subfield><subfield code="c">DE-627</subfield><subfield code="e">rakwb</subfield></datafield><datafield tag="041" ind1=" " ind2=" "><subfield code="a">eng</subfield></datafield><datafield tag="050" ind1=" " ind2="0"><subfield code="a">QC350-467</subfield></datafield><datafield tag="100" ind1="0" ind2=" "><subfield code="a">Yu Yao</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">Decoding of Raman spectroscopy-encoded suspension arrays based on the detail constraint cycle domain adaptive model</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="c">2022</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">Text</subfield><subfield code="b">txt</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">Computermedien</subfield><subfield code="b">c</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">Online-Ressource</subfield><subfield code="b">cr</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">Previous studies have already shown that Raman spectroscopy can be used in the encoding of suspension array technology. However, almost all existing convolutional neural network-based decoding approaches rely on supervision with ground truth, and may not be well generalized to unseen datasets, which were collected under different experimental conditions, applying with the same coded material. In this study, we propose an improved model based on CyCADA, named as Detail constraint Cycle Domain Adaptive Model (DCDA). DCDA implements the classification of unseen datasets through domain adaptation, adapts representations at the encode level with decoder-share, and enforces coding features while leveraging a feat loss. To improve detailed structural constraints, DCDA takes downsample connection and skips connection. Our model improves the poor generalization of existing models and saves the cost of the labeling process for unseen target datasets. Compared with other models, extensive experiments and ablation studies show the superiority of DCDA in terms of classification stability and generalization. The model proposed by the research achieves a classification with an accuracy of 100% when applied in datasets, in which the spectrum in the source domain is far less than the target domain.</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Domain adaption</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">suspension arrays</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">deep learning</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Raman spectrum</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">generalization</subfield></datafield><datafield tag="653" ind1=" " ind2="0"><subfield code="a">Technology</subfield></datafield><datafield tag="653" ind1=" " ind2="0"><subfield code="a">T</subfield></datafield><datafield tag="653" ind1=" " ind2="0"><subfield code="a">Optics. Light</subfield></datafield><datafield tag="700" ind1="0" ind2=" "><subfield code="a">Kaiwen Xue</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="0" ind2=" "><subfield code="a">Liwang Liu</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="0" ind2=" "><subfield code="a">Shanshan Zhu</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="0" ind2=" "><subfield code="a">Chengfeng Yue</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="0" ind2=" "><subfield code="a">Yanhong Ji</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="773" ind1="0" ind2="8"><subfield code="i">In</subfield><subfield code="t">Journal of Innovative Optical Health Sciences</subfield><subfield code="d">World Scientific Publishing, 2017</subfield><subfield code="g">15(2022), 04</subfield><subfield code="w">(DE-627)60940315X</subfield><subfield code="w">(DE-600)2515441-2</subfield><subfield code="x">17937205</subfield><subfield code="7">nnns</subfield></datafield><datafield tag="773" ind1="1" ind2="8"><subfield code="g">volume:15</subfield><subfield code="g">year:2022</subfield><subfield code="g">number:04</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="u">https://doi.org/10.1142/S1793545822500250</subfield><subfield code="z">kostenfrei</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="u">https://doaj.org/article/bb2234f8ac5f4d4592c433baed65fc5c</subfield><subfield code="z">kostenfrei</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="u">https://www.worldscientific.com/doi/10.1142/S1793545822500250</subfield><subfield code="z">kostenfrei</subfield></datafield><datafield tag="856" ind1="4" ind2="2"><subfield code="u">https://doaj.org/toc/1793-5458</subfield><subfield code="y">Journal toc</subfield><subfield code="z">kostenfrei</subfield></datafield><datafield tag="856" ind1="4" ind2="2"><subfield code="u">https://doaj.org/toc/1793-7205</subfield><subfield code="y">Journal toc</subfield><subfield code="z">kostenfrei</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_USEFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SYSFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_DOAJ</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-PHA</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_11</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_20</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_22</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_23</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_24</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_31</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_39</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_40</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_60</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_62</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_63</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_65</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_69</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_70</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_73</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_74</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_95</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_105</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_110</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_151</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_161</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_170</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_206</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_213</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_230</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_285</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_293</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_370</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_602</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_2005</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_2009</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_2011</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_2014</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_2055</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_2111</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4012</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4037</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4112</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4125</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4126</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4249</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4305</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4306</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4307</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4313</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4322</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4323</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4324</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4325</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4335</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4338</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4367</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_4700</subfield></datafield><datafield tag="951" ind1=" " ind2=" "><subfield code="a">AR</subfield></datafield><datafield tag="952" ind1=" " ind2=" "><subfield code="d">15</subfield><subfield code="j">2022</subfield><subfield code="e">04</subfield></datafield></record></collection>
|
score |
7.3991013 |