Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler
This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networ...
Ausführliche Beschreibung
Autor*in: |
Li, Guoqiang [verfasserIn] |
---|
Format: |
E-Artikel |
---|---|
Sprache: |
Englisch |
Erschienen: |
2014transfer abstract |
---|
Schlagwörter: |
---|
Umfang: |
10 |
---|
Übergeordnetes Werk: |
Enthalten in: Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing - 2012, the official journal of the International Neural Network Society, European Neural Network Society and Japanese Neural Network Society, Amsterdam |
---|---|
Übergeordnetes Werk: |
volume:51 ; year:2014 ; pages:57-66 ; extent:10 |
Links: |
---|
DOI / URN: |
10.1016/j.neunet.2013.12.006 |
---|
Katalog-ID: |
ELV034012796 |
---|
LEADER | 01000caa a22002652 4500 | ||
---|---|---|---|
001 | ELV034012796 | ||
003 | DE-627 | ||
005 | 20230625195628.0 | ||
007 | cr uuu---uuuuu | ||
008 | 180603s2014 xx |||||o 00| ||eng c | ||
024 | 7 | |a 10.1016/j.neunet.2013.12.006 |2 doi | |
028 | 5 | 2 | |a GBVA2014015000004.pica |
035 | |a (DE-627)ELV034012796 | ||
035 | |a (ELSEVIER)S0893-6080(13)00300-6 | ||
040 | |a DE-627 |b ger |c DE-627 |e rakwb | ||
041 | |a eng | ||
082 | 0 | |a 004 | |
082 | 0 | 4 | |a 004 |q DE-600 |
082 | 0 | 4 | |a 620 |q VZ |
082 | 0 | 4 | |a 610 |q VZ |
084 | |a 77.50 |2 bkl | ||
100 | 1 | |a Li, Guoqiang |e verfasserin |4 aut | |
245 | 1 | 0 | |a Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler |
264 | 1 | |c 2014transfer abstract | |
300 | |a 10 | ||
336 | |a nicht spezifiziert |b zzz |2 rdacontent | ||
337 | |a nicht spezifiziert |b z |2 rdamedia | ||
338 | |a nicht spezifiziert |b zu |2 rdacarrier | ||
520 | |a This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. | ||
520 | |a This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. | ||
650 | 7 | |a Least square |2 Elsevier | |
650 | 7 | |a Coal-fired boiler |2 Elsevier | |
650 | 7 | |a Artificial neural network |2 Elsevier | |
650 | 7 | |a Least Square Fast Learning Network |2 Elsevier | |
700 | 1 | |a Niu, Peifeng |4 oth | |
700 | 1 | |a Wang, Huaibao |4 oth | |
700 | 1 | |a Liu, Yongchao |4 oth | |
773 | 0 | 8 | |i Enthalten in |n Elsevier |t Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing |d 2012 |d the official journal of the International Neural Network Society, European Neural Network Society and Japanese Neural Network Society |g Amsterdam |w (DE-627)ELV016218965 |
773 | 1 | 8 | |g volume:51 |g year:2014 |g pages:57-66 |g extent:10 |
856 | 4 | 0 | |u https://doi.org/10.1016/j.neunet.2013.12.006 |3 Volltext |
912 | |a GBV_USEFLAG_U | ||
912 | |a GBV_ELV | ||
912 | |a SYSFLAG_U | ||
912 | |a SSG-OLC-PHA | ||
936 | b | k | |a 77.50 |j Psychophysiologie |q VZ |
951 | |a AR | ||
952 | |d 51 |j 2014 |h 57-66 |g 10 | ||
953 | |2 045F |a 004 |
author_variant |
g l gl |
---|---|
matchkey_str |
liguoqiangniupeifengwanghuaibaoliuyongch:2014----:essurfslannntokomdlnteobsinfiin |
hierarchy_sort_str |
2014transfer abstract |
bklnumber |
77.50 |
publishDate |
2014 |
allfields |
10.1016/j.neunet.2013.12.006 doi GBVA2014015000004.pica (DE-627)ELV034012796 (ELSEVIER)S0893-6080(13)00300-6 DE-627 ger DE-627 rakwb eng 004 004 DE-600 620 VZ 610 VZ 77.50 bkl Li, Guoqiang verfasserin aut Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler 2014transfer abstract 10 nicht spezifiziert zzz rdacontent nicht spezifiziert z rdamedia nicht spezifiziert zu rdacarrier This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. Least square Elsevier Coal-fired boiler Elsevier Artificial neural network Elsevier Least Square Fast Learning Network Elsevier Niu, Peifeng oth Wang, Huaibao oth Liu, Yongchao oth Enthalten in Elsevier Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing 2012 the official journal of the International Neural Network Society, European Neural Network Society and Japanese Neural Network Society Amsterdam (DE-627)ELV016218965 volume:51 year:2014 pages:57-66 extent:10 https://doi.org/10.1016/j.neunet.2013.12.006 Volltext GBV_USEFLAG_U GBV_ELV SYSFLAG_U SSG-OLC-PHA 77.50 Psychophysiologie VZ AR 51 2014 57-66 10 045F 004 |
spelling |
10.1016/j.neunet.2013.12.006 doi GBVA2014015000004.pica (DE-627)ELV034012796 (ELSEVIER)S0893-6080(13)00300-6 DE-627 ger DE-627 rakwb eng 004 004 DE-600 620 VZ 610 VZ 77.50 bkl Li, Guoqiang verfasserin aut Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler 2014transfer abstract 10 nicht spezifiziert zzz rdacontent nicht spezifiziert z rdamedia nicht spezifiziert zu rdacarrier This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. Least square Elsevier Coal-fired boiler Elsevier Artificial neural network Elsevier Least Square Fast Learning Network Elsevier Niu, Peifeng oth Wang, Huaibao oth Liu, Yongchao oth Enthalten in Elsevier Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing 2012 the official journal of the International Neural Network Society, European Neural Network Society and Japanese Neural Network Society Amsterdam (DE-627)ELV016218965 volume:51 year:2014 pages:57-66 extent:10 https://doi.org/10.1016/j.neunet.2013.12.006 Volltext GBV_USEFLAG_U GBV_ELV SYSFLAG_U SSG-OLC-PHA 77.50 Psychophysiologie VZ AR 51 2014 57-66 10 045F 004 |
allfields_unstemmed |
10.1016/j.neunet.2013.12.006 doi GBVA2014015000004.pica (DE-627)ELV034012796 (ELSEVIER)S0893-6080(13)00300-6 DE-627 ger DE-627 rakwb eng 004 004 DE-600 620 VZ 610 VZ 77.50 bkl Li, Guoqiang verfasserin aut Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler 2014transfer abstract 10 nicht spezifiziert zzz rdacontent nicht spezifiziert z rdamedia nicht spezifiziert zu rdacarrier This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. Least square Elsevier Coal-fired boiler Elsevier Artificial neural network Elsevier Least Square Fast Learning Network Elsevier Niu, Peifeng oth Wang, Huaibao oth Liu, Yongchao oth Enthalten in Elsevier Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing 2012 the official journal of the International Neural Network Society, European Neural Network Society and Japanese Neural Network Society Amsterdam (DE-627)ELV016218965 volume:51 year:2014 pages:57-66 extent:10 https://doi.org/10.1016/j.neunet.2013.12.006 Volltext GBV_USEFLAG_U GBV_ELV SYSFLAG_U SSG-OLC-PHA 77.50 Psychophysiologie VZ AR 51 2014 57-66 10 045F 004 |
allfieldsGer |
10.1016/j.neunet.2013.12.006 doi GBVA2014015000004.pica (DE-627)ELV034012796 (ELSEVIER)S0893-6080(13)00300-6 DE-627 ger DE-627 rakwb eng 004 004 DE-600 620 VZ 610 VZ 77.50 bkl Li, Guoqiang verfasserin aut Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler 2014transfer abstract 10 nicht spezifiziert zzz rdacontent nicht spezifiziert z rdamedia nicht spezifiziert zu rdacarrier This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. Least square Elsevier Coal-fired boiler Elsevier Artificial neural network Elsevier Least Square Fast Learning Network Elsevier Niu, Peifeng oth Wang, Huaibao oth Liu, Yongchao oth Enthalten in Elsevier Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing 2012 the official journal of the International Neural Network Society, European Neural Network Society and Japanese Neural Network Society Amsterdam (DE-627)ELV016218965 volume:51 year:2014 pages:57-66 extent:10 https://doi.org/10.1016/j.neunet.2013.12.006 Volltext GBV_USEFLAG_U GBV_ELV SYSFLAG_U SSG-OLC-PHA 77.50 Psychophysiologie VZ AR 51 2014 57-66 10 045F 004 |
allfieldsSound |
10.1016/j.neunet.2013.12.006 doi GBVA2014015000004.pica (DE-627)ELV034012796 (ELSEVIER)S0893-6080(13)00300-6 DE-627 ger DE-627 rakwb eng 004 004 DE-600 620 VZ 610 VZ 77.50 bkl Li, Guoqiang verfasserin aut Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler 2014transfer abstract 10 nicht spezifiziert zzz rdacontent nicht spezifiziert z rdamedia nicht spezifiziert zu rdacarrier This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. Least square Elsevier Coal-fired boiler Elsevier Artificial neural network Elsevier Least Square Fast Learning Network Elsevier Niu, Peifeng oth Wang, Huaibao oth Liu, Yongchao oth Enthalten in Elsevier Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing 2012 the official journal of the International Neural Network Society, European Neural Network Society and Japanese Neural Network Society Amsterdam (DE-627)ELV016218965 volume:51 year:2014 pages:57-66 extent:10 https://doi.org/10.1016/j.neunet.2013.12.006 Volltext GBV_USEFLAG_U GBV_ELV SYSFLAG_U SSG-OLC-PHA 77.50 Psychophysiologie VZ AR 51 2014 57-66 10 045F 004 |
language |
English |
source |
Enthalten in Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing Amsterdam volume:51 year:2014 pages:57-66 extent:10 |
sourceStr |
Enthalten in Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing Amsterdam volume:51 year:2014 pages:57-66 extent:10 |
format_phy_str_mv |
Article |
bklname |
Psychophysiologie |
institution |
findex.gbv.de |
topic_facet |
Least square Coal-fired boiler Artificial neural network Least Square Fast Learning Network |
dewey-raw |
004 |
isfreeaccess_bool |
false |
container_title |
Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing |
authorswithroles_txt_mv |
Li, Guoqiang @@aut@@ Niu, Peifeng @@oth@@ Wang, Huaibao @@oth@@ Liu, Yongchao @@oth@@ |
publishDateDaySort_date |
2014-01-01T00:00:00Z |
hierarchy_top_id |
ELV016218965 |
dewey-sort |
14 |
id |
ELV034012796 |
language_de |
englisch |
fullrecord |
<?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01000caa a22002652 4500</leader><controlfield tag="001">ELV034012796</controlfield><controlfield tag="003">DE-627</controlfield><controlfield tag="005">20230625195628.0</controlfield><controlfield tag="007">cr uuu---uuuuu</controlfield><controlfield tag="008">180603s2014 xx |||||o 00| ||eng c</controlfield><datafield tag="024" ind1="7" ind2=" "><subfield code="a">10.1016/j.neunet.2013.12.006</subfield><subfield code="2">doi</subfield></datafield><datafield tag="028" ind1="5" ind2="2"><subfield code="a">GBVA2014015000004.pica</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-627)ELV034012796</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(ELSEVIER)S0893-6080(13)00300-6</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-627</subfield><subfield code="b">ger</subfield><subfield code="c">DE-627</subfield><subfield code="e">rakwb</subfield></datafield><datafield tag="041" ind1=" " ind2=" "><subfield code="a">eng</subfield></datafield><datafield tag="082" ind1="0" ind2=" "><subfield code="a">004</subfield></datafield><datafield tag="082" ind1="0" ind2="4"><subfield code="a">004</subfield><subfield code="q">DE-600</subfield></datafield><datafield tag="082" ind1="0" ind2="4"><subfield code="a">620</subfield><subfield code="q">VZ</subfield></datafield><datafield tag="082" ind1="0" ind2="4"><subfield code="a">610</subfield><subfield code="q">VZ</subfield></datafield><datafield tag="084" ind1=" " ind2=" "><subfield code="a">77.50</subfield><subfield code="2">bkl</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Li, Guoqiang</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="c">2014transfer abstract</subfield></datafield><datafield tag="300" ind1=" " ind2=" "><subfield code="a">10</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">nicht spezifiziert</subfield><subfield code="b">zzz</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">nicht spezifiziert</subfield><subfield code="b">z</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">nicht spezifiziert</subfield><subfield code="b">zu</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed.</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed.</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Least square</subfield><subfield code="2">Elsevier</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Coal-fired boiler</subfield><subfield code="2">Elsevier</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Artificial neural network</subfield><subfield code="2">Elsevier</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Least Square Fast Learning Network</subfield><subfield code="2">Elsevier</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Niu, Peifeng</subfield><subfield code="4">oth</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Wang, Huaibao</subfield><subfield code="4">oth</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Liu, Yongchao</subfield><subfield code="4">oth</subfield></datafield><datafield tag="773" ind1="0" ind2="8"><subfield code="i">Enthalten in</subfield><subfield code="n">Elsevier</subfield><subfield code="t">Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing</subfield><subfield code="d">2012</subfield><subfield code="d">the official journal of the International Neural Network Society, European Neural Network Society and Japanese Neural Network Society</subfield><subfield code="g">Amsterdam</subfield><subfield code="w">(DE-627)ELV016218965</subfield></datafield><datafield tag="773" ind1="1" ind2="8"><subfield code="g">volume:51</subfield><subfield code="g">year:2014</subfield><subfield code="g">pages:57-66</subfield><subfield code="g">extent:10</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="u">https://doi.org/10.1016/j.neunet.2013.12.006</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_USEFLAG_U</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ELV</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SYSFLAG_U</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-PHA</subfield></datafield><datafield tag="936" ind1="b" ind2="k"><subfield code="a">77.50</subfield><subfield code="j">Psychophysiologie</subfield><subfield code="q">VZ</subfield></datafield><datafield tag="951" ind1=" " ind2=" "><subfield code="a">AR</subfield></datafield><datafield tag="952" ind1=" " ind2=" "><subfield code="d">51</subfield><subfield code="j">2014</subfield><subfield code="h">57-66</subfield><subfield code="g">10</subfield></datafield><datafield tag="953" ind1=" " ind2=" "><subfield code="2">045F</subfield><subfield code="a">004</subfield></datafield></record></collection>
|
author |
Li, Guoqiang |
spellingShingle |
Li, Guoqiang ddc 004 ddc 620 ddc 610 bkl 77.50 Elsevier Least square Elsevier Coal-fired boiler Elsevier Artificial neural network Elsevier Least Square Fast Learning Network Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler |
authorStr |
Li, Guoqiang |
ppnlink_with_tag_str_mv |
@@773@@(DE-627)ELV016218965 |
format |
electronic Article |
dewey-ones |
004 - Data processing & computer science 620 - Engineering & allied operations 610 - Medicine & health |
delete_txt_mv |
keep |
author_role |
aut |
collection |
elsevier |
remote_str |
true |
illustrated |
Not Illustrated |
topic_title |
004 004 DE-600 620 VZ 610 VZ 77.50 bkl Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler Least square Elsevier Coal-fired boiler Elsevier Artificial neural network Elsevier Least Square Fast Learning Network Elsevier |
topic |
ddc 004 ddc 620 ddc 610 bkl 77.50 Elsevier Least square Elsevier Coal-fired boiler Elsevier Artificial neural network Elsevier Least Square Fast Learning Network |
topic_unstemmed |
ddc 004 ddc 620 ddc 610 bkl 77.50 Elsevier Least square Elsevier Coal-fired boiler Elsevier Artificial neural network Elsevier Least Square Fast Learning Network |
topic_browse |
ddc 004 ddc 620 ddc 610 bkl 77.50 Elsevier Least square Elsevier Coal-fired boiler Elsevier Artificial neural network Elsevier Least Square Fast Learning Network |
format_facet |
Elektronische Aufsätze Aufsätze Elektronische Ressource |
format_main_str_mv |
Text Zeitschrift/Artikel |
carriertype_str_mv |
zu |
author2_variant |
p n pn h w hw y l yl |
hierarchy_parent_title |
Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing |
hierarchy_parent_id |
ELV016218965 |
dewey-tens |
000 - Computer science, knowledge & systems 620 - Engineering 610 - Medicine & health |
hierarchy_top_title |
Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing |
isfreeaccess_txt |
false |
familylinks_str_mv |
(DE-627)ELV016218965 |
title |
Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler |
ctrlnum |
(DE-627)ELV034012796 (ELSEVIER)S0893-6080(13)00300-6 |
title_full |
Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler |
author_sort |
Li, Guoqiang |
journal |
Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing |
journalStr |
Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing |
lang_code |
eng |
isOA_bool |
false |
dewey-hundreds |
000 - Computer science, information & general works 600 - Technology |
recordtype |
marc |
publishDateSort |
2014 |
contenttype_str_mv |
zzz |
container_start_page |
57 |
author_browse |
Li, Guoqiang |
container_volume |
51 |
physical |
10 |
class |
004 004 DE-600 620 VZ 610 VZ 77.50 bkl |
format_se |
Elektronische Aufsätze |
author-letter |
Li, Guoqiang |
doi_str_mv |
10.1016/j.neunet.2013.12.006 |
dewey-full |
004 620 610 |
title_sort |
least square fast learning network for modeling the combustion efficiency of a 300wm coal-fired boiler |
title_auth |
Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler |
abstract |
This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. |
abstractGer |
This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. |
abstract_unstemmed |
This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed. |
collection_details |
GBV_USEFLAG_U GBV_ELV SYSFLAG_U SSG-OLC-PHA |
title_short |
Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler |
url |
https://doi.org/10.1016/j.neunet.2013.12.006 |
remote_bool |
true |
author2 |
Niu, Peifeng Wang, Huaibao Liu, Yongchao |
author2Str |
Niu, Peifeng Wang, Huaibao Liu, Yongchao |
ppnlink |
ELV016218965 |
mediatype_str_mv |
z |
isOA_txt |
false |
hochschulschrift_bool |
false |
author2_role |
oth oth oth |
doi_str |
10.1016/j.neunet.2013.12.006 |
up_date |
2024-07-06T20:02:47.984Z |
_version_ |
1803861274639990784 |
fullrecord_marcxml |
<?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01000caa a22002652 4500</leader><controlfield tag="001">ELV034012796</controlfield><controlfield tag="003">DE-627</controlfield><controlfield tag="005">20230625195628.0</controlfield><controlfield tag="007">cr uuu---uuuuu</controlfield><controlfield tag="008">180603s2014 xx |||||o 00| ||eng c</controlfield><datafield tag="024" ind1="7" ind2=" "><subfield code="a">10.1016/j.neunet.2013.12.006</subfield><subfield code="2">doi</subfield></datafield><datafield tag="028" ind1="5" ind2="2"><subfield code="a">GBVA2014015000004.pica</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-627)ELV034012796</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(ELSEVIER)S0893-6080(13)00300-6</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-627</subfield><subfield code="b">ger</subfield><subfield code="c">DE-627</subfield><subfield code="e">rakwb</subfield></datafield><datafield tag="041" ind1=" " ind2=" "><subfield code="a">eng</subfield></datafield><datafield tag="082" ind1="0" ind2=" "><subfield code="a">004</subfield></datafield><datafield tag="082" ind1="0" ind2="4"><subfield code="a">004</subfield><subfield code="q">DE-600</subfield></datafield><datafield tag="082" ind1="0" ind2="4"><subfield code="a">620</subfield><subfield code="q">VZ</subfield></datafield><datafield tag="082" ind1="0" ind2="4"><subfield code="a">610</subfield><subfield code="q">VZ</subfield></datafield><datafield tag="084" ind1=" " ind2=" "><subfield code="a">77.50</subfield><subfield code="2">bkl</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Li, Guoqiang</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="c">2014transfer abstract</subfield></datafield><datafield tag="300" ind1=" " ind2=" "><subfield code="a">10</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">nicht spezifiziert</subfield><subfield code="b">zzz</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">nicht spezifiziert</subfield><subfield code="b">z</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">nicht spezifiziert</subfield><subfield code="b">zu</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed.</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">This paper presents a novel artificial neural network with a very fast learning speed, all of whose weights and biases are determined by the twice Least Square method, so it is called Least Square Fast Learning Network (LSFLN). In addition, there is another difference from conventional neural networks, which is that the output neurons of LSFLN not only receive the information from the hidden layer neurons, but also receive the external information itself directly from the input neurons. In order to test the validity of LSFLN, it is applied to 6 classical regression applications, and also employed to build the functional relation between the combustion efficiency and operating parameters of a 300WM coal-fired boiler. Experimental results show that, compared with other methods, LSFLN with very less hidden neurons could achieve much better regression precision and generalization ability at a much faster learning speed.</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Least square</subfield><subfield code="2">Elsevier</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Coal-fired boiler</subfield><subfield code="2">Elsevier</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Artificial neural network</subfield><subfield code="2">Elsevier</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Least Square Fast Learning Network</subfield><subfield code="2">Elsevier</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Niu, Peifeng</subfield><subfield code="4">oth</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Wang, Huaibao</subfield><subfield code="4">oth</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Liu, Yongchao</subfield><subfield code="4">oth</subfield></datafield><datafield tag="773" ind1="0" ind2="8"><subfield code="i">Enthalten in</subfield><subfield code="n">Elsevier</subfield><subfield code="t">Regulatory design for RES-E support mechanisms: Learning curves, market structure, and burden-sharing</subfield><subfield code="d">2012</subfield><subfield code="d">the official journal of the International Neural Network Society, European Neural Network Society and Japanese Neural Network Society</subfield><subfield code="g">Amsterdam</subfield><subfield code="w">(DE-627)ELV016218965</subfield></datafield><datafield tag="773" ind1="1" ind2="8"><subfield code="g">volume:51</subfield><subfield code="g">year:2014</subfield><subfield code="g">pages:57-66</subfield><subfield code="g">extent:10</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="u">https://doi.org/10.1016/j.neunet.2013.12.006</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_USEFLAG_U</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ELV</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SYSFLAG_U</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-PHA</subfield></datafield><datafield tag="936" ind1="b" ind2="k"><subfield code="a">77.50</subfield><subfield code="j">Psychophysiologie</subfield><subfield code="q">VZ</subfield></datafield><datafield tag="951" ind1=" " ind2=" "><subfield code="a">AR</subfield></datafield><datafield tag="952" ind1=" " ind2=" "><subfield code="d">51</subfield><subfield code="j">2014</subfield><subfield code="h">57-66</subfield><subfield code="g">10</subfield></datafield><datafield tag="953" ind1=" " ind2=" "><subfield code="2">045F</subfield><subfield code="a">004</subfield></datafield></record></collection>
|
score |
7.401394 |