Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/29876
Full metadata record
DC FieldValueLanguage
dc.contributor.authorYannakakis, Georgios N.-
dc.contributor.authorLevine, John-
dc.contributor.authorHallam, John-
dc.date.accessioned2018-05-09T06:37:37Z-
dc.date.available2018-05-09T06:37:37Z-
dc.date.issued2007-
dc.identifier.citationYannakakis, G. N., Levine, J., & Hallam, J. (2007). Emerging cooperation with minimal effort: rewarding over mimicking. IEEE Transactions on Evolutionary Computation, 11(3), 382-396.en_GB
dc.identifier.urihttps://www.um.edu.mt/library/oar//handle/123456789/29876-
dc.description.abstractThis paper compares supervised and unsupervised learning mechanisms for the emergence of cooperative multiagent spatial coordination using a top-down approach. By observing the global performance of a group of homogeneous agents—supported by a nonglobal knowledge of their environment—we attempt to extract information about the minimum size of the agent neurocontroller and the type of learning mechanism that collectively generate high-performing and robust behaviors with minimal computational effort. Consequently, a methodology for obtaining controllers of minimal size is introduced and a comparative study between supervised and unsupervised learning mechanisms for the generation of successful collective behaviors is presented. We have developed a prototype simulated world for our studies. This case study is primarily a computer games inspired world but its main features are also biologically plausible. The two specific tasks that the agents are tested in are the competing strategies of obstacle-avoidance and target-achievement. We demonstrate that cooperative behavior among agents, which is supported only by limited communication, appears to be necessary for the problem’s efficient solution and that learning by rewarding the behavior of agent groups constitutes a more efficient and computationally preferred generic approach than supervised learning approaches in such complex multiagent worlds.en_GB
dc.language.isoenen_GB
dc.publisherInstitute of Electrical and Electronics Engineers Inc.en_GB
dc.rightsinfo:eu-repo/semantics/restrictedAccessen_GB
dc.subjectComputer games -- Case studiesen_GB
dc.subjectGenetic algorithmsen_GB
dc.subjectMachine learningen_GB
dc.subjectNeural networks (Computer science)en_GB
dc.titleEmerging cooperation with minimal effort : rewarding over mimickingen_GB
dc.typearticleen_GB
dc.rights.holderThe copyright of this work belongs to the author(s)/publisher. The rights of this work are as defined by the appropriate Copyright Legislation or as modified by any successive legislation. Users may access this work and can make use of the information contained in accordance with the Copyright Legislation provided that the author must be properly acknowledged. Further distribution or reproduction in any format is prohibited without the prior permission of the copyright holder.en_GB
dc.description.reviewedpeer-revieweden_GB
dc.identifier.doi10.1109/TEVC.2006.882429-
dc.publication.titleIEEE Transactions on Evolutionary Computationen_GB
Appears in Collections:Scholarly Works - InsDG

Files in This Item:
File Description SizeFormat 
Emerging_cooperation_with_minimal_effort_rewarding_over_mimicking_2007.pdf
  Restricted Access
857.98 kBAdobe PDFView/Open Request a copy


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.