The analysis and performance evaluation of the pheromone-Q-learning algorithm
Abstract: The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a com...
Ausführliche Beschreibung
Autor*in: |
Monekosso, N. [verfasserIn] Remagnino, P. [verfasserIn] |
---|
Format: |
E-Artikel |
---|
Erschienen: |
Oxford, UK: Blackwell Publishing ; 2004 |
---|
Schlagwörter: |
---|
Umfang: |
Online-Ressource |
---|
Reproduktion: |
2004 ; Blackwell Publishing Journal Backfiles 1879-2005 |
---|---|
Übergeordnetes Werk: |
In: Expert systems - Oxford [u.a.] : Wiley-Blackwell, 1997, 21(2004), 2, Seite 0 |
Übergeordnetes Werk: |
volume:21 ; year:2004 ; number:2 ; pages:0 |
Links: |
---|
DOI / URN: |
10.1111/j.1468-0394.2004.00265.x |
---|
Katalog-ID: |
NLEJ242374905 |
---|
LEADER | 01000caa a22002652 4500 | ||
---|---|---|---|
001 | NLEJ242374905 | ||
003 | DE-627 | ||
005 | 20210707154130.0 | ||
007 | cr uuu---uuuuu | ||
008 | 120427s2004 xx |||||o 00| ||und c | ||
024 | 7 | |a 10.1111/j.1468-0394.2004.00265.x |2 doi | |
035 | |a (DE-627)NLEJ242374905 | ||
040 | |a DE-627 |b ger |c DE-627 |e rakwb | ||
100 | 1 | |a Monekosso, N. |e verfasserin |4 aut | |
245 | 1 | 0 | |a The analysis and performance evaluation of the pheromone-Q-learning algorithm |
264 | 1 | |a Oxford, UK |b Blackwell Publishing |c 2004 | |
300 | |a Online-Ressource | ||
336 | |a nicht spezifiziert |b zzz |2 rdacontent | ||
337 | |a nicht spezifiziert |b z |2 rdamedia | ||
338 | |a nicht spezifiziert |b zu |2 rdacarrier | ||
520 | |a Abstract: The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone. | ||
533 | |d 2004 |f Blackwell Publishing Journal Backfiles 1879-2005 |7 |2004|||||||||| | ||
650 | 4 | |a multi-agent systems | |
700 | 1 | |a Remagnino, P. |e verfasserin |4 aut | |
773 | 0 | 8 | |i In |t Expert systems |d Oxford [u.a.] : Wiley-Blackwell, 1997 |g 21(2004), 2, Seite 0 |h Online-Ressource |w (DE-627)NLEJ243925662 |w (DE-600)2016958-9 |x 1468-0394 |7 nnns |
773 | 1 | 8 | |g volume:21 |g year:2004 |g number:2 |g pages:0 |
856 | 4 | 0 | |u http://dx.doi.org/10.1111/j.1468-0394.2004.00265.x |q text/html |x Verlag |z Deutschlandweit zugänglich |3 Volltext |
912 | |a GBV_USEFLAG_U | ||
912 | |a ZDB-1-DJB | ||
912 | |a GBV_NL_ARTICLE | ||
951 | |a AR | ||
952 | |d 21 |j 2004 |e 2 |h 0 |
author_variant |
n m nm p r pr |
---|---|
matchkey_str |
article:14680394:2004----::haayiadefraceautootehrmn |
hierarchy_sort_str |
2004 |
publishDate |
2004 |
allfields |
10.1111/j.1468-0394.2004.00265.x doi (DE-627)NLEJ242374905 DE-627 ger DE-627 rakwb Monekosso, N. verfasserin aut The analysis and performance evaluation of the pheromone-Q-learning algorithm Oxford, UK Blackwell Publishing 2004 Online-Ressource nicht spezifiziert zzz rdacontent nicht spezifiziert z rdamedia nicht spezifiziert zu rdacarrier Abstract: The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone. 2004 Blackwell Publishing Journal Backfiles 1879-2005 |2004|||||||||| multi-agent systems Remagnino, P. verfasserin aut In Expert systems Oxford [u.a.] : Wiley-Blackwell, 1997 21(2004), 2, Seite 0 Online-Ressource (DE-627)NLEJ243925662 (DE-600)2016958-9 1468-0394 nnns volume:21 year:2004 number:2 pages:0 http://dx.doi.org/10.1111/j.1468-0394.2004.00265.x text/html Verlag Deutschlandweit zugänglich Volltext GBV_USEFLAG_U ZDB-1-DJB GBV_NL_ARTICLE AR 21 2004 2 0 |
spelling |
10.1111/j.1468-0394.2004.00265.x doi (DE-627)NLEJ242374905 DE-627 ger DE-627 rakwb Monekosso, N. verfasserin aut The analysis and performance evaluation of the pheromone-Q-learning algorithm Oxford, UK Blackwell Publishing 2004 Online-Ressource nicht spezifiziert zzz rdacontent nicht spezifiziert z rdamedia nicht spezifiziert zu rdacarrier Abstract: The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone. 2004 Blackwell Publishing Journal Backfiles 1879-2005 |2004|||||||||| multi-agent systems Remagnino, P. verfasserin aut In Expert systems Oxford [u.a.] : Wiley-Blackwell, 1997 21(2004), 2, Seite 0 Online-Ressource (DE-627)NLEJ243925662 (DE-600)2016958-9 1468-0394 nnns volume:21 year:2004 number:2 pages:0 http://dx.doi.org/10.1111/j.1468-0394.2004.00265.x text/html Verlag Deutschlandweit zugänglich Volltext GBV_USEFLAG_U ZDB-1-DJB GBV_NL_ARTICLE AR 21 2004 2 0 |
allfields_unstemmed |
10.1111/j.1468-0394.2004.00265.x doi (DE-627)NLEJ242374905 DE-627 ger DE-627 rakwb Monekosso, N. verfasserin aut The analysis and performance evaluation of the pheromone-Q-learning algorithm Oxford, UK Blackwell Publishing 2004 Online-Ressource nicht spezifiziert zzz rdacontent nicht spezifiziert z rdamedia nicht spezifiziert zu rdacarrier Abstract: The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone. 2004 Blackwell Publishing Journal Backfiles 1879-2005 |2004|||||||||| multi-agent systems Remagnino, P. verfasserin aut In Expert systems Oxford [u.a.] : Wiley-Blackwell, 1997 21(2004), 2, Seite 0 Online-Ressource (DE-627)NLEJ243925662 (DE-600)2016958-9 1468-0394 nnns volume:21 year:2004 number:2 pages:0 http://dx.doi.org/10.1111/j.1468-0394.2004.00265.x text/html Verlag Deutschlandweit zugänglich Volltext GBV_USEFLAG_U ZDB-1-DJB GBV_NL_ARTICLE AR 21 2004 2 0 |
allfieldsGer |
10.1111/j.1468-0394.2004.00265.x doi (DE-627)NLEJ242374905 DE-627 ger DE-627 rakwb Monekosso, N. verfasserin aut The analysis and performance evaluation of the pheromone-Q-learning algorithm Oxford, UK Blackwell Publishing 2004 Online-Ressource nicht spezifiziert zzz rdacontent nicht spezifiziert z rdamedia nicht spezifiziert zu rdacarrier Abstract: The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone. 2004 Blackwell Publishing Journal Backfiles 1879-2005 |2004|||||||||| multi-agent systems Remagnino, P. verfasserin aut In Expert systems Oxford [u.a.] : Wiley-Blackwell, 1997 21(2004), 2, Seite 0 Online-Ressource (DE-627)NLEJ243925662 (DE-600)2016958-9 1468-0394 nnns volume:21 year:2004 number:2 pages:0 http://dx.doi.org/10.1111/j.1468-0394.2004.00265.x text/html Verlag Deutschlandweit zugänglich Volltext GBV_USEFLAG_U ZDB-1-DJB GBV_NL_ARTICLE AR 21 2004 2 0 |
allfieldsSound |
10.1111/j.1468-0394.2004.00265.x doi (DE-627)NLEJ242374905 DE-627 ger DE-627 rakwb Monekosso, N. verfasserin aut The analysis and performance evaluation of the pheromone-Q-learning algorithm Oxford, UK Blackwell Publishing 2004 Online-Ressource nicht spezifiziert zzz rdacontent nicht spezifiziert z rdamedia nicht spezifiziert zu rdacarrier Abstract: The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone. 2004 Blackwell Publishing Journal Backfiles 1879-2005 |2004|||||||||| multi-agent systems Remagnino, P. verfasserin aut In Expert systems Oxford [u.a.] : Wiley-Blackwell, 1997 21(2004), 2, Seite 0 Online-Ressource (DE-627)NLEJ243925662 (DE-600)2016958-9 1468-0394 nnns volume:21 year:2004 number:2 pages:0 http://dx.doi.org/10.1111/j.1468-0394.2004.00265.x text/html Verlag Deutschlandweit zugänglich Volltext GBV_USEFLAG_U ZDB-1-DJB GBV_NL_ARTICLE AR 21 2004 2 0 |
source |
In Expert systems 21(2004), 2, Seite 0 volume:21 year:2004 number:2 pages:0 |
sourceStr |
In Expert systems 21(2004), 2, Seite 0 volume:21 year:2004 number:2 pages:0 |
format_phy_str_mv |
Article |
institution |
findex.gbv.de |
topic_facet |
multi-agent systems |
isfreeaccess_bool |
false |
container_title |
Expert systems |
authorswithroles_txt_mv |
Monekosso, N. @@aut@@ Remagnino, P. @@aut@@ |
publishDateDaySort_date |
2004-01-01T00:00:00Z |
hierarchy_top_id |
NLEJ243925662 |
id |
NLEJ242374905 |
fullrecord |
<?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01000caa a22002652 4500</leader><controlfield tag="001">NLEJ242374905</controlfield><controlfield tag="003">DE-627</controlfield><controlfield tag="005">20210707154130.0</controlfield><controlfield tag="007">cr uuu---uuuuu</controlfield><controlfield tag="008">120427s2004 xx |||||o 00| ||und c</controlfield><datafield tag="024" ind1="7" ind2=" "><subfield code="a">10.1111/j.1468-0394.2004.00265.x</subfield><subfield code="2">doi</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-627)NLEJ242374905</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-627</subfield><subfield code="b">ger</subfield><subfield code="c">DE-627</subfield><subfield code="e">rakwb</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Monekosso, N.</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">The analysis and performance evaluation of the pheromone-Q-learning algorithm</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="a">Oxford, UK</subfield><subfield code="b">Blackwell Publishing</subfield><subfield code="c">2004</subfield></datafield><datafield tag="300" ind1=" " ind2=" "><subfield code="a">Online-Ressource</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">nicht spezifiziert</subfield><subfield code="b">zzz</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">nicht spezifiziert</subfield><subfield code="b">z</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">nicht spezifiziert</subfield><subfield code="b">zu</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">Abstract: The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone.</subfield></datafield><datafield tag="533" ind1=" " ind2=" "><subfield code="d">2004</subfield><subfield code="f">Blackwell Publishing Journal Backfiles 1879-2005</subfield><subfield code="7">|2004||||||||||</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">multi-agent systems</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Remagnino, P.</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="773" ind1="0" ind2="8"><subfield code="i">In</subfield><subfield code="t">Expert systems</subfield><subfield code="d">Oxford [u.a.] : Wiley-Blackwell, 1997</subfield><subfield code="g">21(2004), 2, Seite 0</subfield><subfield code="h">Online-Ressource</subfield><subfield code="w">(DE-627)NLEJ243925662</subfield><subfield code="w">(DE-600)2016958-9</subfield><subfield code="x">1468-0394</subfield><subfield code="7">nnns</subfield></datafield><datafield tag="773" ind1="1" ind2="8"><subfield code="g">volume:21</subfield><subfield code="g">year:2004</subfield><subfield code="g">number:2</subfield><subfield code="g">pages:0</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="u">http://dx.doi.org/10.1111/j.1468-0394.2004.00265.x</subfield><subfield code="q">text/html</subfield><subfield code="x">Verlag</subfield><subfield code="z">Deutschlandweit zugänglich</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_USEFLAG_U</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">ZDB-1-DJB</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_NL_ARTICLE</subfield></datafield><datafield tag="951" ind1=" " ind2=" "><subfield code="a">AR</subfield></datafield><datafield tag="952" ind1=" " ind2=" "><subfield code="d">21</subfield><subfield code="j">2004</subfield><subfield code="e">2</subfield><subfield code="h">0</subfield></datafield></record></collection>
|
series2 |
Blackwell Publishing Journal Backfiles 1879-2005 |
author |
Monekosso, N. |
spellingShingle |
Monekosso, N. misc multi-agent systems The analysis and performance evaluation of the pheromone-Q-learning algorithm |
authorStr |
Monekosso, N. |
ppnlink_with_tag_str_mv |
@@773@@(DE-627)NLEJ243925662 |
format |
electronic Article |
delete_txt_mv |
keep |
author_role |
aut aut |
collection |
NL |
publishPlace |
Oxford, UK |
remote_str |
true |
illustrated |
Not Illustrated |
issn |
1468-0394 |
topic_title |
The analysis and performance evaluation of the pheromone-Q-learning algorithm multi-agent systems |
publisher |
Blackwell Publishing |
publisherStr |
Blackwell Publishing |
topic |
misc multi-agent systems |
topic_unstemmed |
misc multi-agent systems |
topic_browse |
misc multi-agent systems |
format_facet |
Elektronische Aufsätze Aufsätze Elektronische Ressource |
format_main_str_mv |
Text Zeitschrift/Artikel |
carriertype_str_mv |
zu |
hierarchy_parent_title |
Expert systems |
hierarchy_parent_id |
NLEJ243925662 |
hierarchy_top_title |
Expert systems |
isfreeaccess_txt |
false |
familylinks_str_mv |
(DE-627)NLEJ243925662 (DE-600)2016958-9 |
title |
The analysis and performance evaluation of the pheromone-Q-learning algorithm |
ctrlnum |
(DE-627)NLEJ242374905 |
title_full |
The analysis and performance evaluation of the pheromone-Q-learning algorithm |
author_sort |
Monekosso, N. |
journal |
Expert systems |
journalStr |
Expert systems |
isOA_bool |
false |
recordtype |
marc |
publishDateSort |
2004 |
contenttype_str_mv |
zzz |
container_start_page |
0 |
author_browse |
Monekosso, N. Remagnino, P. |
container_volume |
21 |
physical |
Online-Ressource |
format_se |
Elektronische Aufsätze |
author-letter |
Monekosso, N. |
doi_str_mv |
10.1111/j.1468-0394.2004.00265.x |
author2-role |
verfasserin |
title_sort |
the analysis and performance evaluation of the pheromone-q-learning algorithm |
title_auth |
The analysis and performance evaluation of the pheromone-Q-learning algorithm |
abstract |
Abstract: The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone. |
abstractGer |
Abstract: The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone. |
abstract_unstemmed |
Abstract: The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone. |
collection_details |
GBV_USEFLAG_U ZDB-1-DJB GBV_NL_ARTICLE |
container_issue |
2 |
title_short |
The analysis and performance evaluation of the pheromone-Q-learning algorithm |
url |
http://dx.doi.org/10.1111/j.1468-0394.2004.00265.x |
remote_bool |
true |
author2 |
Remagnino, P. |
author2Str |
Remagnino, P. |
ppnlink |
NLEJ243925662 |
mediatype_str_mv |
z |
isOA_txt |
false |
hochschulschrift_bool |
false |
doi_str |
10.1111/j.1468-0394.2004.00265.x |
up_date |
2024-07-06T01:45:43.630Z |
_version_ |
1803792252803219456 |
fullrecord_marcxml |
<?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01000caa a22002652 4500</leader><controlfield tag="001">NLEJ242374905</controlfield><controlfield tag="003">DE-627</controlfield><controlfield tag="005">20210707154130.0</controlfield><controlfield tag="007">cr uuu---uuuuu</controlfield><controlfield tag="008">120427s2004 xx |||||o 00| ||und c</controlfield><datafield tag="024" ind1="7" ind2=" "><subfield code="a">10.1111/j.1468-0394.2004.00265.x</subfield><subfield code="2">doi</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-627)NLEJ242374905</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-627</subfield><subfield code="b">ger</subfield><subfield code="c">DE-627</subfield><subfield code="e">rakwb</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Monekosso, N.</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">The analysis and performance evaluation of the pheromone-Q-learning algorithm</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="a">Oxford, UK</subfield><subfield code="b">Blackwell Publishing</subfield><subfield code="c">2004</subfield></datafield><datafield tag="300" ind1=" " ind2=" "><subfield code="a">Online-Ressource</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">nicht spezifiziert</subfield><subfield code="b">zzz</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">nicht spezifiziert</subfield><subfield code="b">z</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">nicht spezifiziert</subfield><subfield code="b">zu</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">Abstract: The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone.</subfield></datafield><datafield tag="533" ind1=" " ind2=" "><subfield code="d">2004</subfield><subfield code="f">Blackwell Publishing Journal Backfiles 1879-2005</subfield><subfield code="7">|2004||||||||||</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">multi-agent systems</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Remagnino, P.</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="773" ind1="0" ind2="8"><subfield code="i">In</subfield><subfield code="t">Expert systems</subfield><subfield code="d">Oxford [u.a.] : Wiley-Blackwell, 1997</subfield><subfield code="g">21(2004), 2, Seite 0</subfield><subfield code="h">Online-Ressource</subfield><subfield code="w">(DE-627)NLEJ243925662</subfield><subfield code="w">(DE-600)2016958-9</subfield><subfield code="x">1468-0394</subfield><subfield code="7">nnns</subfield></datafield><datafield tag="773" ind1="1" ind2="8"><subfield code="g">volume:21</subfield><subfield code="g">year:2004</subfield><subfield code="g">number:2</subfield><subfield code="g">pages:0</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="u">http://dx.doi.org/10.1111/j.1468-0394.2004.00265.x</subfield><subfield code="q">text/html</subfield><subfield code="x">Verlag</subfield><subfield code="z">Deutschlandweit zugänglich</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_USEFLAG_U</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">ZDB-1-DJB</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_NL_ARTICLE</subfield></datafield><datafield tag="951" ind1=" " ind2=" "><subfield code="a">AR</subfield></datafield><datafield tag="952" ind1=" " ind2=" "><subfield code="d">21</subfield><subfield code="j">2004</subfield><subfield code="e">2</subfield><subfield code="h">0</subfield></datafield></record></collection>
|
score |
7.4008837 |