Speeding Task Allocation Search for Reconfigurations in Adaptive Distributed Embedded Systems Using Deep Reinforcement Learning.

Uloženo v:
Podrobná bibliografie
Název: Speeding Task Allocation Search for Reconfigurations in Adaptive Distributed Embedded Systems Using Deep Reinforcement Learning.
Autoři: Rotaeche, Ramón, Ballesteros, Alberto, Proenza, Julián
Zdroj: Sensors (14248220); Jan2023, Vol. 23 Issue 1, p548, 23p
Témata: POLYNOMIAL time algorithms, REINFORCEMENT learning, COMBINATORIAL optimization, TIME management
Abstrakt: A Critical Adaptive Distributed Embedded System (CADES) is a group of interconnected nodes that must carry out a set of tasks to achieve a common goal, while fulfilling several requirements associated with their critical (e.g., hard real-time requirements) and adaptive nature. In these systems, a key challenge is to solve, in a timely manner, the combinatorial optimization problem involved in finding the best way to allocate the tasks to the available nodes (i.e., the task allocation) taking into account aspects such as the computational costs of the tasks and the computational capacity of the nodes. This problem is not trivial and there is no known polynomial time algorithm to find the optimal solution. Several studies have proposed Deep Reinforcement Learning (DRL) approaches to solve combinatorial optimization problems and, in this work, we explore the application of such approaches to the task allocation problem in CADESs. We first discuss the potential advantages of using a DRL-based approach over several heuristic-based approaches to allocate tasks in CADESs and we then demonstrate how a DRL-based approach can achieve similar results for the best performing heuristic in terms of optimality of the allocation, while requiring less time to generate such allocation. [ABSTRACT FROM AUTHOR]
Copyright of Sensors (14248220) is the property of MDPI and its content may not be copied or emailed to multiple sites without the copyright holder's express written permission. Additionally, content may not be used with any artificial intelligence tools or machine learning technologies. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
Databáze: Complementary Index
FullText Text:
  Availability: 0
CustomLinks:
  – Url: http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=search&db=pmc&term=1424-8220[TA]+AND+548[PG]+AND+2023[PDAT]
    Name: FREE - PubMed Central (ISSN based link)
    Category: fullText
    Text: Full Text
    Icon: https://imageserver.ebscohost.com/NetImages/iconPdf.gif
    MouseOverText: Check this PubMed for the article full text.
  – Url: https://resolver.ebscohost.com/openurl?sid=EBSCO:edb&genre=article&issn=14248220&ISBN=&volume=23&issue=1&date=20230101&spage=548&pages=548-570&title=Sensors (14248220)&atitle=Speeding%20Task%20Allocation%20Search%20for%20Reconfigurations%20in%20Adaptive%20Distributed%20Embedded%20Systems%20Using%20Deep%20Reinforcement%20Learning.&aulast=Rotaeche%2C%20Ram%C3%B3n&id=DOI:10.3390/s23010548
    Name: Full Text Finder
    Category: fullText
    Text: Full Text Finder
    Icon: https://imageserver.ebscohost.com/branding/images/FTF.gif
    MouseOverText: Full Text Finder
  – Url: https://www.webofscience.com/api/gateway?GWVersion=2&SrcApp=EBSCO&SrcAuth=EBSCO&DestApp=WOS&ServiceName=TransferToWoS&DestLinkType=GeneralSearchSummary&Func=Links&author=Rotaeche%20R
    Name: ISI
    Category: fullText
    Text: Nájsť tento článok vo Web of Science
    Icon: https://imagesrvr.epnet.com/ls/20docs.gif
    MouseOverText: Nájsť tento článok vo Web of Science
Header DbId: edb
DbLabel: Complementary Index
An: 161186259
RelevancyScore: 938
AccessLevel: 6
PubType: Academic Journal
PubTypeId: academicJournal
PreciseRelevancyScore: 937.849487304688
IllustrationInfo
Items – Name: Title
  Label: Title
  Group: Ti
  Data: Speeding Task Allocation Search for Reconfigurations in Adaptive Distributed Embedded Systems Using Deep Reinforcement Learning.
– Name: Author
  Label: Authors
  Group: Au
  Data: <searchLink fieldCode="AR" term="%22Rotaeche%2C+Ramón%22">Rotaeche, Ramón</searchLink><br /><searchLink fieldCode="AR" term="%22Ballesteros%2C+Alberto%22">Ballesteros, Alberto</searchLink><br /><searchLink fieldCode="AR" term="%22Proenza%2C+Julián%22">Proenza, Julián</searchLink>
– Name: TitleSource
  Label: Source
  Group: Src
  Data: Sensors (14248220); Jan2023, Vol. 23 Issue 1, p548, 23p
– Name: Subject
  Label: Subject Terms
  Group: Su
  Data: <searchLink fieldCode="DE" term="%22POLYNOMIAL+time+algorithms%22">POLYNOMIAL time algorithms</searchLink><br /><searchLink fieldCode="DE" term="%22REINFORCEMENT+learning%22">REINFORCEMENT learning</searchLink><br /><searchLink fieldCode="DE" term="%22COMBINATORIAL+optimization%22">COMBINATORIAL optimization</searchLink><br /><searchLink fieldCode="DE" term="%22TIME+management%22">TIME management</searchLink>
– Name: Abstract
  Label: Abstract
  Group: Ab
  Data: A Critical Adaptive Distributed Embedded System (CADES) is a group of interconnected nodes that must carry out a set of tasks to achieve a common goal, while fulfilling several requirements associated with their critical (e.g., hard real-time requirements) and adaptive nature. In these systems, a key challenge is to solve, in a timely manner, the combinatorial optimization problem involved in finding the best way to allocate the tasks to the available nodes (i.e., the task allocation) taking into account aspects such as the computational costs of the tasks and the computational capacity of the nodes. This problem is not trivial and there is no known polynomial time algorithm to find the optimal solution. Several studies have proposed Deep Reinforcement Learning (DRL) approaches to solve combinatorial optimization problems and, in this work, we explore the application of such approaches to the task allocation problem in CADESs. We first discuss the potential advantages of using a DRL-based approach over several heuristic-based approaches to allocate tasks in CADESs and we then demonstrate how a DRL-based approach can achieve similar results for the best performing heuristic in terms of optimality of the allocation, while requiring less time to generate such allocation. [ABSTRACT FROM AUTHOR]
– Name: Abstract
  Label:
  Group: Ab
  Data: <i>Copyright of Sensors (14248220) is the property of MDPI and its content may not be copied or emailed to multiple sites without the copyright holder's express written permission. Additionally, content may not be used with any artificial intelligence tools or machine learning technologies. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract.</i> (Copyright applies to all Abstracts.)
PLink https://erproxy.cvtisr.sk/sfx/access?url=https://search.ebscohost.com/login.aspx?direct=true&site=eds-live&db=edb&AN=161186259
RecordInfo BibRecord:
  BibEntity:
    Identifiers:
      – Type: doi
        Value: 10.3390/s23010548
    Languages:
      – Code: eng
        Text: English
    PhysicalDescription:
      Pagination:
        PageCount: 23
        StartPage: 548
    Subjects:
      – SubjectFull: POLYNOMIAL time algorithms
        Type: general
      – SubjectFull: REINFORCEMENT learning
        Type: general
      – SubjectFull: COMBINATORIAL optimization
        Type: general
      – SubjectFull: TIME management
        Type: general
    Titles:
      – TitleFull: Speeding Task Allocation Search for Reconfigurations in Adaptive Distributed Embedded Systems Using Deep Reinforcement Learning.
        Type: main
  BibRelationships:
    HasContributorRelationships:
      – PersonEntity:
          Name:
            NameFull: Rotaeche, Ramón
      – PersonEntity:
          Name:
            NameFull: Ballesteros, Alberto
      – PersonEntity:
          Name:
            NameFull: Proenza, Julián
    IsPartOfRelationships:
      – BibEntity:
          Dates:
            – D: 01
              M: 01
              Text: Jan2023
              Type: published
              Y: 2023
          Identifiers:
            – Type: issn-print
              Value: 14248220
          Numbering:
            – Type: volume
              Value: 23
            – Type: issue
              Value: 1
          Titles:
            – TitleFull: Sensors (14248220)
              Type: main
ResultId 1