CF-DAML: Distributed automated machine learning based on collaborative filtering
Abstract The search for a good machine learning (ML) model takes a long time and requires the considerations of many alternatives, including data preprocessing, algorithm selection, and hyperparameter tuning methods. Thus, tedious searches face a combinatorial explosion problem. In this work, we bui...
Ausführliche Beschreibung
Autor*in: |
Liu, Pengjie [verfasserIn] |
---|
Format: |
Artikel |
---|---|
Sprache: |
Englisch |
Erschienen: |
2022 |
---|
Schlagwörter: |
---|
Anmerkung: |
© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022 |
---|
Übergeordnetes Werk: |
Enthalten in: Applied intelligence - Springer US, 1991, 52(2022), 15 vom: 31. März, Seite 17145-17169 |
---|---|
Übergeordnetes Werk: |
volume:52 ; year:2022 ; number:15 ; day:31 ; month:03 ; pages:17145-17169 |
Links: |
---|
DOI / URN: |
10.1007/s10489-021-03049-z |
---|
Katalog-ID: |
OLC2080016865 |
---|
LEADER | 01000caa a22002652 4500 | ||
---|---|---|---|
001 | OLC2080016865 | ||
003 | DE-627 | ||
005 | 20230506090825.0 | ||
007 | tu | ||
008 | 230131s2022 xx ||||| 00| ||eng c | ||
024 | 7 | |a 10.1007/s10489-021-03049-z |2 doi | |
035 | |a (DE-627)OLC2080016865 | ||
035 | |a (DE-He213)s10489-021-03049-z-p | ||
040 | |a DE-627 |b ger |c DE-627 |e rakwb | ||
041 | |a eng | ||
082 | 0 | 4 | |a 004 |q VZ |
100 | 1 | |a Liu, Pengjie |e verfasserin |4 aut | |
245 | 1 | 0 | |a CF-DAML: Distributed automated machine learning based on collaborative filtering |
264 | 1 | |c 2022 | |
336 | |a Text |b txt |2 rdacontent | ||
337 | |a ohne Hilfsmittel zu benutzen |b n |2 rdamedia | ||
338 | |a Band |b nc |2 rdacarrier | ||
500 | |a © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022 | ||
520 | |a Abstract The search for a good machine learning (ML) model takes a long time and requires the considerations of many alternatives, including data preprocessing, algorithm selection, and hyperparameter tuning methods. Thus, tedious searches face a combinatorial explosion problem. In this work, we build a new automated machine learning (AutoML) system called CF-DAML, a distributed automated system based on collaborative filtering (CF), to address these challenges by recommending and training suitable models for supervised learning tasks. CF-DAML first computes some informative meta-features for a new dataset, then uses a weighted $$l_1$$-norm (W1-norm) to accurately calculate the k nearest neighbors (kNN) of the new dataset, and finally recommends the top N models with good performances on each of its neighbors to the new dataset. We also design a distributed system (DSTM) for training the models to reduce the time complexity substantially. In addition, we develop a multilayer selective stacked ensemble system (MSSE), whose base models are selected from among suitable candidate models based on their runtimes, classification accuracies, and diversities, to enhance the stability of CF-DAML. To our knowledge, this is the first work to combine memory-based CF and the selective stacked ensemble to solve the AutoML problem. Extensive experiments are conducted on many UCI datasets and the comparative results demonstrate that our approach outperforms the current state-of-the-art methods. | ||
650 | 4 | |a Automated machine learning | |
650 | 4 | |a Collaborative filtering | |
650 | 4 | |a Weighted | |
650 | 4 | |a -norm | |
650 | 4 | |a Distributed automated system | |
650 | 4 | |a Multilayer selective stacked ensemble | |
700 | 1 | |a Pan, Fucheng |4 aut | |
700 | 1 | |a Zhou, Xiaofeng |0 (orcid)0000-0001-9837-1261 |4 aut | |
700 | 1 | |a Li, Shuai |4 aut | |
700 | 1 | |a Jin, Liang |4 aut | |
773 | 0 | 8 | |i Enthalten in |t Applied intelligence |d Springer US, 1991 |g 52(2022), 15 vom: 31. März, Seite 17145-17169 |w (DE-627)130990515 |w (DE-600)1080229-0 |w (DE-576)029154286 |x 0924-669X |7 nnns |
773 | 1 | 8 | |g volume:52 |g year:2022 |g number:15 |g day:31 |g month:03 |g pages:17145-17169 |
856 | 4 | 1 | |u https://doi.org/10.1007/s10489-021-03049-z |z lizenzpflichtig |3 Volltext |
912 | |a GBV_USEFLAG_A | ||
912 | |a SYSFLAG_A | ||
912 | |a GBV_OLC | ||
912 | |a SSG-OLC-MAT | ||
951 | |a AR | ||
952 | |d 52 |j 2022 |e 15 |b 31 |c 03 |h 17145-17169 |
author_variant |
p l pl f p fp x z xz s l sl l j lj |
---|---|
matchkey_str |
article:0924669X:2022----::falitiueatmtdahnlannbsdno |
hierarchy_sort_str |
2022 |
publishDate |
2022 |
allfields |
10.1007/s10489-021-03049-z doi (DE-627)OLC2080016865 (DE-He213)s10489-021-03049-z-p DE-627 ger DE-627 rakwb eng 004 VZ Liu, Pengjie verfasserin aut CF-DAML: Distributed automated machine learning based on collaborative filtering 2022 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022 Abstract The search for a good machine learning (ML) model takes a long time and requires the considerations of many alternatives, including data preprocessing, algorithm selection, and hyperparameter tuning methods. Thus, tedious searches face a combinatorial explosion problem. In this work, we build a new automated machine learning (AutoML) system called CF-DAML, a distributed automated system based on collaborative filtering (CF), to address these challenges by recommending and training suitable models for supervised learning tasks. CF-DAML first computes some informative meta-features for a new dataset, then uses a weighted $$l_1$$-norm (W1-norm) to accurately calculate the k nearest neighbors (kNN) of the new dataset, and finally recommends the top N models with good performances on each of its neighbors to the new dataset. We also design a distributed system (DSTM) for training the models to reduce the time complexity substantially. In addition, we develop a multilayer selective stacked ensemble system (MSSE), whose base models are selected from among suitable candidate models based on their runtimes, classification accuracies, and diversities, to enhance the stability of CF-DAML. To our knowledge, this is the first work to combine memory-based CF and the selective stacked ensemble to solve the AutoML problem. Extensive experiments are conducted on many UCI datasets and the comparative results demonstrate that our approach outperforms the current state-of-the-art methods. Automated machine learning Collaborative filtering Weighted -norm Distributed automated system Multilayer selective stacked ensemble Pan, Fucheng aut Zhou, Xiaofeng (orcid)0000-0001-9837-1261 aut Li, Shuai aut Jin, Liang aut Enthalten in Applied intelligence Springer US, 1991 52(2022), 15 vom: 31. März, Seite 17145-17169 (DE-627)130990515 (DE-600)1080229-0 (DE-576)029154286 0924-669X nnns volume:52 year:2022 number:15 day:31 month:03 pages:17145-17169 https://doi.org/10.1007/s10489-021-03049-z lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT AR 52 2022 15 31 03 17145-17169 |
spelling |
10.1007/s10489-021-03049-z doi (DE-627)OLC2080016865 (DE-He213)s10489-021-03049-z-p DE-627 ger DE-627 rakwb eng 004 VZ Liu, Pengjie verfasserin aut CF-DAML: Distributed automated machine learning based on collaborative filtering 2022 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022 Abstract The search for a good machine learning (ML) model takes a long time and requires the considerations of many alternatives, including data preprocessing, algorithm selection, and hyperparameter tuning methods. Thus, tedious searches face a combinatorial explosion problem. In this work, we build a new automated machine learning (AutoML) system called CF-DAML, a distributed automated system based on collaborative filtering (CF), to address these challenges by recommending and training suitable models for supervised learning tasks. CF-DAML first computes some informative meta-features for a new dataset, then uses a weighted $$l_1$$-norm (W1-norm) to accurately calculate the k nearest neighbors (kNN) of the new dataset, and finally recommends the top N models with good performances on each of its neighbors to the new dataset. We also design a distributed system (DSTM) for training the models to reduce the time complexity substantially. In addition, we develop a multilayer selective stacked ensemble system (MSSE), whose base models are selected from among suitable candidate models based on their runtimes, classification accuracies, and diversities, to enhance the stability of CF-DAML. To our knowledge, this is the first work to combine memory-based CF and the selective stacked ensemble to solve the AutoML problem. Extensive experiments are conducted on many UCI datasets and the comparative results demonstrate that our approach outperforms the current state-of-the-art methods. Automated machine learning Collaborative filtering Weighted -norm Distributed automated system Multilayer selective stacked ensemble Pan, Fucheng aut Zhou, Xiaofeng (orcid)0000-0001-9837-1261 aut Li, Shuai aut Jin, Liang aut Enthalten in Applied intelligence Springer US, 1991 52(2022), 15 vom: 31. März, Seite 17145-17169 (DE-627)130990515 (DE-600)1080229-0 (DE-576)029154286 0924-669X nnns volume:52 year:2022 number:15 day:31 month:03 pages:17145-17169 https://doi.org/10.1007/s10489-021-03049-z lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT AR 52 2022 15 31 03 17145-17169 |
allfields_unstemmed |
10.1007/s10489-021-03049-z doi (DE-627)OLC2080016865 (DE-He213)s10489-021-03049-z-p DE-627 ger DE-627 rakwb eng 004 VZ Liu, Pengjie verfasserin aut CF-DAML: Distributed automated machine learning based on collaborative filtering 2022 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022 Abstract The search for a good machine learning (ML) model takes a long time and requires the considerations of many alternatives, including data preprocessing, algorithm selection, and hyperparameter tuning methods. Thus, tedious searches face a combinatorial explosion problem. In this work, we build a new automated machine learning (AutoML) system called CF-DAML, a distributed automated system based on collaborative filtering (CF), to address these challenges by recommending and training suitable models for supervised learning tasks. CF-DAML first computes some informative meta-features for a new dataset, then uses a weighted $$l_1$$-norm (W1-norm) to accurately calculate the k nearest neighbors (kNN) of the new dataset, and finally recommends the top N models with good performances on each of its neighbors to the new dataset. We also design a distributed system (DSTM) for training the models to reduce the time complexity substantially. In addition, we develop a multilayer selective stacked ensemble system (MSSE), whose base models are selected from among suitable candidate models based on their runtimes, classification accuracies, and diversities, to enhance the stability of CF-DAML. To our knowledge, this is the first work to combine memory-based CF and the selective stacked ensemble to solve the AutoML problem. Extensive experiments are conducted on many UCI datasets and the comparative results demonstrate that our approach outperforms the current state-of-the-art methods. Automated machine learning Collaborative filtering Weighted -norm Distributed automated system Multilayer selective stacked ensemble Pan, Fucheng aut Zhou, Xiaofeng (orcid)0000-0001-9837-1261 aut Li, Shuai aut Jin, Liang aut Enthalten in Applied intelligence Springer US, 1991 52(2022), 15 vom: 31. März, Seite 17145-17169 (DE-627)130990515 (DE-600)1080229-0 (DE-576)029154286 0924-669X nnns volume:52 year:2022 number:15 day:31 month:03 pages:17145-17169 https://doi.org/10.1007/s10489-021-03049-z lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT AR 52 2022 15 31 03 17145-17169 |
allfieldsGer |
10.1007/s10489-021-03049-z doi (DE-627)OLC2080016865 (DE-He213)s10489-021-03049-z-p DE-627 ger DE-627 rakwb eng 004 VZ Liu, Pengjie verfasserin aut CF-DAML: Distributed automated machine learning based on collaborative filtering 2022 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022 Abstract The search for a good machine learning (ML) model takes a long time and requires the considerations of many alternatives, including data preprocessing, algorithm selection, and hyperparameter tuning methods. Thus, tedious searches face a combinatorial explosion problem. In this work, we build a new automated machine learning (AutoML) system called CF-DAML, a distributed automated system based on collaborative filtering (CF), to address these challenges by recommending and training suitable models for supervised learning tasks. CF-DAML first computes some informative meta-features for a new dataset, then uses a weighted $$l_1$$-norm (W1-norm) to accurately calculate the k nearest neighbors (kNN) of the new dataset, and finally recommends the top N models with good performances on each of its neighbors to the new dataset. We also design a distributed system (DSTM) for training the models to reduce the time complexity substantially. In addition, we develop a multilayer selective stacked ensemble system (MSSE), whose base models are selected from among suitable candidate models based on their runtimes, classification accuracies, and diversities, to enhance the stability of CF-DAML. To our knowledge, this is the first work to combine memory-based CF and the selective stacked ensemble to solve the AutoML problem. Extensive experiments are conducted on many UCI datasets and the comparative results demonstrate that our approach outperforms the current state-of-the-art methods. Automated machine learning Collaborative filtering Weighted -norm Distributed automated system Multilayer selective stacked ensemble Pan, Fucheng aut Zhou, Xiaofeng (orcid)0000-0001-9837-1261 aut Li, Shuai aut Jin, Liang aut Enthalten in Applied intelligence Springer US, 1991 52(2022), 15 vom: 31. März, Seite 17145-17169 (DE-627)130990515 (DE-600)1080229-0 (DE-576)029154286 0924-669X nnns volume:52 year:2022 number:15 day:31 month:03 pages:17145-17169 https://doi.org/10.1007/s10489-021-03049-z lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT AR 52 2022 15 31 03 17145-17169 |
allfieldsSound |
10.1007/s10489-021-03049-z doi (DE-627)OLC2080016865 (DE-He213)s10489-021-03049-z-p DE-627 ger DE-627 rakwb eng 004 VZ Liu, Pengjie verfasserin aut CF-DAML: Distributed automated machine learning based on collaborative filtering 2022 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022 Abstract The search for a good machine learning (ML) model takes a long time and requires the considerations of many alternatives, including data preprocessing, algorithm selection, and hyperparameter tuning methods. Thus, tedious searches face a combinatorial explosion problem. In this work, we build a new automated machine learning (AutoML) system called CF-DAML, a distributed automated system based on collaborative filtering (CF), to address these challenges by recommending and training suitable models for supervised learning tasks. CF-DAML first computes some informative meta-features for a new dataset, then uses a weighted $$l_1$$-norm (W1-norm) to accurately calculate the k nearest neighbors (kNN) of the new dataset, and finally recommends the top N models with good performances on each of its neighbors to the new dataset. We also design a distributed system (DSTM) for training the models to reduce the time complexity substantially. In addition, we develop a multilayer selective stacked ensemble system (MSSE), whose base models are selected from among suitable candidate models based on their runtimes, classification accuracies, and diversities, to enhance the stability of CF-DAML. To our knowledge, this is the first work to combine memory-based CF and the selective stacked ensemble to solve the AutoML problem. Extensive experiments are conducted on many UCI datasets and the comparative results demonstrate that our approach outperforms the current state-of-the-art methods. Automated machine learning Collaborative filtering Weighted -norm Distributed automated system Multilayer selective stacked ensemble Pan, Fucheng aut Zhou, Xiaofeng (orcid)0000-0001-9837-1261 aut Li, Shuai aut Jin, Liang aut Enthalten in Applied intelligence Springer US, 1991 52(2022), 15 vom: 31. März, Seite 17145-17169 (DE-627)130990515 (DE-600)1080229-0 (DE-576)029154286 0924-669X nnns volume:52 year:2022 number:15 day:31 month:03 pages:17145-17169 https://doi.org/10.1007/s10489-021-03049-z lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT AR 52 2022 15 31 03 17145-17169 |
language |
English |
source |
Enthalten in Applied intelligence 52(2022), 15 vom: 31. März, Seite 17145-17169 volume:52 year:2022 number:15 day:31 month:03 pages:17145-17169 |
sourceStr |
Enthalten in Applied intelligence 52(2022), 15 vom: 31. März, Seite 17145-17169 volume:52 year:2022 number:15 day:31 month:03 pages:17145-17169 |
format_phy_str_mv |
Article |
institution |
findex.gbv.de |
topic_facet |
Automated machine learning Collaborative filtering Weighted -norm Distributed automated system Multilayer selective stacked ensemble |
dewey-raw |
004 |
isfreeaccess_bool |
false |
container_title |
Applied intelligence |
authorswithroles_txt_mv |
Liu, Pengjie @@aut@@ Pan, Fucheng @@aut@@ Zhou, Xiaofeng @@aut@@ Li, Shuai @@aut@@ Jin, Liang @@aut@@ |
publishDateDaySort_date |
2022-03-31T00:00:00Z |
hierarchy_top_id |
130990515 |
dewey-sort |
14 |
id |
OLC2080016865 |
language_de |
englisch |
fullrecord |
<?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01000caa a22002652 4500</leader><controlfield tag="001">OLC2080016865</controlfield><controlfield tag="003">DE-627</controlfield><controlfield tag="005">20230506090825.0</controlfield><controlfield tag="007">tu</controlfield><controlfield tag="008">230131s2022 xx ||||| 00| ||eng c</controlfield><datafield tag="024" ind1="7" ind2=" "><subfield code="a">10.1007/s10489-021-03049-z</subfield><subfield code="2">doi</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-627)OLC2080016865</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-He213)s10489-021-03049-z-p</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-627</subfield><subfield code="b">ger</subfield><subfield code="c">DE-627</subfield><subfield code="e">rakwb</subfield></datafield><datafield tag="041" ind1=" " ind2=" "><subfield code="a">eng</subfield></datafield><datafield tag="082" ind1="0" ind2="4"><subfield code="a">004</subfield><subfield code="q">VZ</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Liu, Pengjie</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">CF-DAML: Distributed automated machine learning based on collaborative filtering</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="c">2022</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">Text</subfield><subfield code="b">txt</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">ohne Hilfsmittel zu benutzen</subfield><subfield code="b">n</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">Band</subfield><subfield code="b">nc</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a">© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">Abstract The search for a good machine learning (ML) model takes a long time and requires the considerations of many alternatives, including data preprocessing, algorithm selection, and hyperparameter tuning methods. Thus, tedious searches face a combinatorial explosion problem. In this work, we build a new automated machine learning (AutoML) system called CF-DAML, a distributed automated system based on collaborative filtering (CF), to address these challenges by recommending and training suitable models for supervised learning tasks. CF-DAML first computes some informative meta-features for a new dataset, then uses a weighted $$l_1$$-norm (W1-norm) to accurately calculate the k nearest neighbors (kNN) of the new dataset, and finally recommends the top N models with good performances on each of its neighbors to the new dataset. We also design a distributed system (DSTM) for training the models to reduce the time complexity substantially. In addition, we develop a multilayer selective stacked ensemble system (MSSE), whose base models are selected from among suitable candidate models based on their runtimes, classification accuracies, and diversities, to enhance the stability of CF-DAML. To our knowledge, this is the first work to combine memory-based CF and the selective stacked ensemble to solve the AutoML problem. Extensive experiments are conducted on many UCI datasets and the comparative results demonstrate that our approach outperforms the current state-of-the-art methods.</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Automated machine learning</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Collaborative filtering</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Weighted</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">-norm</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Distributed automated system</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Multilayer selective stacked ensemble</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Pan, Fucheng</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Zhou, Xiaofeng</subfield><subfield code="0">(orcid)0000-0001-9837-1261</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Li, Shuai</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Jin, Liang</subfield><subfield code="4">aut</subfield></datafield><datafield tag="773" ind1="0" ind2="8"><subfield code="i">Enthalten in</subfield><subfield code="t">Applied intelligence</subfield><subfield code="d">Springer US, 1991</subfield><subfield code="g">52(2022), 15 vom: 31. März, Seite 17145-17169</subfield><subfield code="w">(DE-627)130990515</subfield><subfield code="w">(DE-600)1080229-0</subfield><subfield code="w">(DE-576)029154286</subfield><subfield code="x">0924-669X</subfield><subfield code="7">nnns</subfield></datafield><datafield tag="773" ind1="1" ind2="8"><subfield code="g">volume:52</subfield><subfield code="g">year:2022</subfield><subfield code="g">number:15</subfield><subfield code="g">day:31</subfield><subfield code="g">month:03</subfield><subfield code="g">pages:17145-17169</subfield></datafield><datafield tag="856" ind1="4" ind2="1"><subfield code="u">https://doi.org/10.1007/s10489-021-03049-z</subfield><subfield code="z">lizenzpflichtig</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_USEFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SYSFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_OLC</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-MAT</subfield></datafield><datafield tag="951" ind1=" " ind2=" "><subfield code="a">AR</subfield></datafield><datafield tag="952" ind1=" " ind2=" "><subfield code="d">52</subfield><subfield code="j">2022</subfield><subfield code="e">15</subfield><subfield code="b">31</subfield><subfield code="c">03</subfield><subfield code="h">17145-17169</subfield></datafield></record></collection>
|
author |
Liu, Pengjie |
spellingShingle |
Liu, Pengjie ddc 004 misc Automated machine learning misc Collaborative filtering misc Weighted misc -norm misc Distributed automated system misc Multilayer selective stacked ensemble CF-DAML: Distributed automated machine learning based on collaborative filtering |
authorStr |
Liu, Pengjie |
ppnlink_with_tag_str_mv |
@@773@@(DE-627)130990515 |
format |
Article |
dewey-ones |
004 - Data processing & computer science |
delete_txt_mv |
keep |
author_role |
aut aut aut aut aut |
collection |
OLC |
remote_str |
false |
illustrated |
Not Illustrated |
issn |
0924-669X |
topic_title |
004 VZ CF-DAML: Distributed automated machine learning based on collaborative filtering Automated machine learning Collaborative filtering Weighted -norm Distributed automated system Multilayer selective stacked ensemble |
topic |
ddc 004 misc Automated machine learning misc Collaborative filtering misc Weighted misc -norm misc Distributed automated system misc Multilayer selective stacked ensemble |
topic_unstemmed |
ddc 004 misc Automated machine learning misc Collaborative filtering misc Weighted misc -norm misc Distributed automated system misc Multilayer selective stacked ensemble |
topic_browse |
ddc 004 misc Automated machine learning misc Collaborative filtering misc Weighted misc -norm misc Distributed automated system misc Multilayer selective stacked ensemble |
format_facet |
Aufsätze Gedruckte Aufsätze |
format_main_str_mv |
Text Zeitschrift/Artikel |
carriertype_str_mv |
nc |
hierarchy_parent_title |
Applied intelligence |
hierarchy_parent_id |
130990515 |
dewey-tens |
000 - Computer science, knowledge & systems |
hierarchy_top_title |
Applied intelligence |
isfreeaccess_txt |
false |
familylinks_str_mv |
(DE-627)130990515 (DE-600)1080229-0 (DE-576)029154286 |
title |
CF-DAML: Distributed automated machine learning based on collaborative filtering |
ctrlnum |
(DE-627)OLC2080016865 (DE-He213)s10489-021-03049-z-p |
title_full |
CF-DAML: Distributed automated machine learning based on collaborative filtering |
author_sort |
Liu, Pengjie |
journal |
Applied intelligence |
journalStr |
Applied intelligence |
lang_code |
eng |
isOA_bool |
false |
dewey-hundreds |
000 - Computer science, information & general works |
recordtype |
marc |
publishDateSort |
2022 |
contenttype_str_mv |
txt |
container_start_page |
17145 |
author_browse |
Liu, Pengjie Pan, Fucheng Zhou, Xiaofeng Li, Shuai Jin, Liang |
container_volume |
52 |
class |
004 VZ |
format_se |
Aufsätze |
author-letter |
Liu, Pengjie |
doi_str_mv |
10.1007/s10489-021-03049-z |
normlink |
(ORCID)0000-0001-9837-1261 |
normlink_prefix_str_mv |
(orcid)0000-0001-9837-1261 |
dewey-full |
004 |
title_sort |
cf-daml: distributed automated machine learning based on collaborative filtering |
title_auth |
CF-DAML: Distributed automated machine learning based on collaborative filtering |
abstract |
Abstract The search for a good machine learning (ML) model takes a long time and requires the considerations of many alternatives, including data preprocessing, algorithm selection, and hyperparameter tuning methods. Thus, tedious searches face a combinatorial explosion problem. In this work, we build a new automated machine learning (AutoML) system called CF-DAML, a distributed automated system based on collaborative filtering (CF), to address these challenges by recommending and training suitable models for supervised learning tasks. CF-DAML first computes some informative meta-features for a new dataset, then uses a weighted $$l_1$$-norm (W1-norm) to accurately calculate the k nearest neighbors (kNN) of the new dataset, and finally recommends the top N models with good performances on each of its neighbors to the new dataset. We also design a distributed system (DSTM) for training the models to reduce the time complexity substantially. In addition, we develop a multilayer selective stacked ensemble system (MSSE), whose base models are selected from among suitable candidate models based on their runtimes, classification accuracies, and diversities, to enhance the stability of CF-DAML. To our knowledge, this is the first work to combine memory-based CF and the selective stacked ensemble to solve the AutoML problem. Extensive experiments are conducted on many UCI datasets and the comparative results demonstrate that our approach outperforms the current state-of-the-art methods. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022 |
abstractGer |
Abstract The search for a good machine learning (ML) model takes a long time and requires the considerations of many alternatives, including data preprocessing, algorithm selection, and hyperparameter tuning methods. Thus, tedious searches face a combinatorial explosion problem. In this work, we build a new automated machine learning (AutoML) system called CF-DAML, a distributed automated system based on collaborative filtering (CF), to address these challenges by recommending and training suitable models for supervised learning tasks. CF-DAML first computes some informative meta-features for a new dataset, then uses a weighted $$l_1$$-norm (W1-norm) to accurately calculate the k nearest neighbors (kNN) of the new dataset, and finally recommends the top N models with good performances on each of its neighbors to the new dataset. We also design a distributed system (DSTM) for training the models to reduce the time complexity substantially. In addition, we develop a multilayer selective stacked ensemble system (MSSE), whose base models are selected from among suitable candidate models based on their runtimes, classification accuracies, and diversities, to enhance the stability of CF-DAML. To our knowledge, this is the first work to combine memory-based CF and the selective stacked ensemble to solve the AutoML problem. Extensive experiments are conducted on many UCI datasets and the comparative results demonstrate that our approach outperforms the current state-of-the-art methods. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022 |
abstract_unstemmed |
Abstract The search for a good machine learning (ML) model takes a long time and requires the considerations of many alternatives, including data preprocessing, algorithm selection, and hyperparameter tuning methods. Thus, tedious searches face a combinatorial explosion problem. In this work, we build a new automated machine learning (AutoML) system called CF-DAML, a distributed automated system based on collaborative filtering (CF), to address these challenges by recommending and training suitable models for supervised learning tasks. CF-DAML first computes some informative meta-features for a new dataset, then uses a weighted $$l_1$$-norm (W1-norm) to accurately calculate the k nearest neighbors (kNN) of the new dataset, and finally recommends the top N models with good performances on each of its neighbors to the new dataset. We also design a distributed system (DSTM) for training the models to reduce the time complexity substantially. In addition, we develop a multilayer selective stacked ensemble system (MSSE), whose base models are selected from among suitable candidate models based on their runtimes, classification accuracies, and diversities, to enhance the stability of CF-DAML. To our knowledge, this is the first work to combine memory-based CF and the selective stacked ensemble to solve the AutoML problem. Extensive experiments are conducted on many UCI datasets and the comparative results demonstrate that our approach outperforms the current state-of-the-art methods. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022 |
collection_details |
GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT |
container_issue |
15 |
title_short |
CF-DAML: Distributed automated machine learning based on collaborative filtering |
url |
https://doi.org/10.1007/s10489-021-03049-z |
remote_bool |
false |
author2 |
Pan, Fucheng Zhou, Xiaofeng Li, Shuai Jin, Liang |
author2Str |
Pan, Fucheng Zhou, Xiaofeng Li, Shuai Jin, Liang |
ppnlink |
130990515 |
mediatype_str_mv |
n |
isOA_txt |
false |
hochschulschrift_bool |
false |
doi_str |
10.1007/s10489-021-03049-z |
up_date |
2024-07-04T02:41:44.314Z |
_version_ |
1803614582808248321 |
fullrecord_marcxml |
<?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01000caa a22002652 4500</leader><controlfield tag="001">OLC2080016865</controlfield><controlfield tag="003">DE-627</controlfield><controlfield tag="005">20230506090825.0</controlfield><controlfield tag="007">tu</controlfield><controlfield tag="008">230131s2022 xx ||||| 00| ||eng c</controlfield><datafield tag="024" ind1="7" ind2=" "><subfield code="a">10.1007/s10489-021-03049-z</subfield><subfield code="2">doi</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-627)OLC2080016865</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-He213)s10489-021-03049-z-p</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-627</subfield><subfield code="b">ger</subfield><subfield code="c">DE-627</subfield><subfield code="e">rakwb</subfield></datafield><datafield tag="041" ind1=" " ind2=" "><subfield code="a">eng</subfield></datafield><datafield tag="082" ind1="0" ind2="4"><subfield code="a">004</subfield><subfield code="q">VZ</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Liu, Pengjie</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">CF-DAML: Distributed automated machine learning based on collaborative filtering</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="c">2022</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">Text</subfield><subfield code="b">txt</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">ohne Hilfsmittel zu benutzen</subfield><subfield code="b">n</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">Band</subfield><subfield code="b">nc</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a">© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">Abstract The search for a good machine learning (ML) model takes a long time and requires the considerations of many alternatives, including data preprocessing, algorithm selection, and hyperparameter tuning methods. Thus, tedious searches face a combinatorial explosion problem. In this work, we build a new automated machine learning (AutoML) system called CF-DAML, a distributed automated system based on collaborative filtering (CF), to address these challenges by recommending and training suitable models for supervised learning tasks. CF-DAML first computes some informative meta-features for a new dataset, then uses a weighted $$l_1$$-norm (W1-norm) to accurately calculate the k nearest neighbors (kNN) of the new dataset, and finally recommends the top N models with good performances on each of its neighbors to the new dataset. We also design a distributed system (DSTM) for training the models to reduce the time complexity substantially. In addition, we develop a multilayer selective stacked ensemble system (MSSE), whose base models are selected from among suitable candidate models based on their runtimes, classification accuracies, and diversities, to enhance the stability of CF-DAML. To our knowledge, this is the first work to combine memory-based CF and the selective stacked ensemble to solve the AutoML problem. Extensive experiments are conducted on many UCI datasets and the comparative results demonstrate that our approach outperforms the current state-of-the-art methods.</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Automated machine learning</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Collaborative filtering</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Weighted</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">-norm</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Distributed automated system</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Multilayer selective stacked ensemble</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Pan, Fucheng</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Zhou, Xiaofeng</subfield><subfield code="0">(orcid)0000-0001-9837-1261</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Li, Shuai</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Jin, Liang</subfield><subfield code="4">aut</subfield></datafield><datafield tag="773" ind1="0" ind2="8"><subfield code="i">Enthalten in</subfield><subfield code="t">Applied intelligence</subfield><subfield code="d">Springer US, 1991</subfield><subfield code="g">52(2022), 15 vom: 31. März, Seite 17145-17169</subfield><subfield code="w">(DE-627)130990515</subfield><subfield code="w">(DE-600)1080229-0</subfield><subfield code="w">(DE-576)029154286</subfield><subfield code="x">0924-669X</subfield><subfield code="7">nnns</subfield></datafield><datafield tag="773" ind1="1" ind2="8"><subfield code="g">volume:52</subfield><subfield code="g">year:2022</subfield><subfield code="g">number:15</subfield><subfield code="g">day:31</subfield><subfield code="g">month:03</subfield><subfield code="g">pages:17145-17169</subfield></datafield><datafield tag="856" ind1="4" ind2="1"><subfield code="u">https://doi.org/10.1007/s10489-021-03049-z</subfield><subfield code="z">lizenzpflichtig</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_USEFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SYSFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_OLC</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-MAT</subfield></datafield><datafield tag="951" ind1=" " ind2=" "><subfield code="a">AR</subfield></datafield><datafield tag="952" ind1=" " ind2=" "><subfield code="d">52</subfield><subfield code="j">2022</subfield><subfield code="e">15</subfield><subfield code="b">31</subfield><subfield code="c">03</subfield><subfield code="h">17145-17169</subfield></datafield></record></collection>
|
score |
7.39931 |