Using Behavioural Programming with Solver, Context, and Deep Reinforcement Learning for Playing a Simplified RoboCup-Type Game
We describe four scenario-based implementations of controllers for a player in a simplified RoboCup-type game. All four implementations are based on the behavioural programming (BP) approach. We first describe a simple controller for the player using the state-of-the-art BPjs tool and then show how...
Saved in:
| Published in: | 2019 ACM/IEEE 22nd International Conference on Model Driven Engineering Languages and Systems Companion (MODELS-C) pp. 243 - 251 |
|---|---|
| Main Authors: | , , , |
| Format: | Conference Proceeding |
| Language: | English |
| Published: |
IEEE
01.09.2019
|
| Subjects: | |
| Online Access: | Get full text |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| Abstract | We describe four scenario-based implementations of controllers for a player in a simplified RoboCup-type game. All four implementations are based on the behavioural programming (BP) approach. We first describe a simple controller for the player using the state-of-the-art BPjs tool and then show how it can be extended in various ways. The first extension is based on a version of BP where the Z3 SMT solver is used to provide mechanisms for richer composition of modules within the BP model. This allows for modules with higher cohesion and lower coupling. It also allows incrementality: we could use the scenarios we developed for the challenge of MDETOOLS'18 and extend the model to handle the new system. The second extension of BP demonstrated in this paper is a set of idioms for subjecting model components to context. One of the differences between this year's challenge and the challenge we dealt with last year is that following the ball is not the only task that a player needs to handle, there is much more to care for. We demonstrate how we used the idioms for handling context to parametrize scenarios like "go to a target" in a dynamic and natural fashion such that modelers can efficiently specify reusable components similar to the way modern user manuals for advanced products are written. Lastly, in an attempt to make the instructions to the robot even more natural, we demonstrate a third extension based on deep reinforcement learning. Towards substantiating the observation that it is easier to explain things to an intelligent agent than to dumb compiler, we demonstrate how the combination of BP and deep reinforcement learning (DRL) allows for giving abstract instructions to the robot and for teaching it to follow them after a short training session. |
|---|---|
| AbstractList | We describe four scenario-based implementations of controllers for a player in a simplified RoboCup-type game. All four implementations are based on the behavioural programming (BP) approach. We first describe a simple controller for the player using the state-of-the-art BPjs tool and then show how it can be extended in various ways. The first extension is based on a version of BP where the Z3 SMT solver is used to provide mechanisms for richer composition of modules within the BP model. This allows for modules with higher cohesion and lower coupling. It also allows incrementality: we could use the scenarios we developed for the challenge of MDETOOLS'18 and extend the model to handle the new system. The second extension of BP demonstrated in this paper is a set of idioms for subjecting model components to context. One of the differences between this year's challenge and the challenge we dealt with last year is that following the ball is not the only task that a player needs to handle, there is much more to care for. We demonstrate how we used the idioms for handling context to parametrize scenarios like "go to a target" in a dynamic and natural fashion such that modelers can efficiently specify reusable components similar to the way modern user manuals for advanced products are written. Lastly, in an attempt to make the instructions to the robot even more natural, we demonstrate a third extension based on deep reinforcement learning. Towards substantiating the observation that it is easier to explain things to an intelligent agent than to dumb compiler, we demonstrate how the combination of BP and deep reinforcement learning (DRL) allows for giving abstract instructions to the robot and for teaching it to follow them after a short training session. |
| Author | Elyasaf, Achiya Sadon, Aviran Weiss, Gera Yaacov, Tom |
| Author_xml | – sequence: 1 givenname: Achiya surname: Elyasaf fullname: Elyasaf, Achiya organization: Ben Gurion University of The Negev – sequence: 2 givenname: Aviran surname: Sadon fullname: Sadon, Aviran organization: Ben Gurion University of The Negev – sequence: 3 givenname: Gera surname: Weiss fullname: Weiss, Gera organization: Ben Gurion University of The Negev – sequence: 4 givenname: Tom surname: Yaacov fullname: Yaacov, Tom organization: Ben Gurion University of The Negev |
| BookMark | eNotjF1PwjAYhWuiiYr8Am76Axj27T5oL3UgmsxA-Lgm3fYWmmzt0g10N_52IXp1Tp485zySW-ssEjICNgFg8vlzOZtnmyCdcAZywhgL5Q0ZyqmAKRcQA4_ZPRm2rclZLKIk5DE8kJ9da-yBvuJRnY07eVXRlXcHr-r6yr9Md6QbV53Rj2nqbIff3ZgqW9IZYkPXaKx2vsAabUczVN5eVxdEV5Xqr13RjambymiDJV273KWnJtj2DdKFqvGJ3GlVtTj8zwHZvc236XuQLRcf6UsWKC7CLghBykJIwRAQQPI816yMI15I0CAZTlGLi5GUpRaRzGNdlEXOyyTSiicx6HBARn-_BhH3jTe18v1eSBaJhIe_1kphgA |
| CODEN | IEEPAD |
| ContentType | Conference Proceeding |
| DBID | 6IE 6IL CBEJK RIE RIL |
| DOI | 10.1109/MODELS-C.2019.00039 |
| DatabaseName | IEEE Electronic Library (IEL) Conference Proceedings IEEE Xplore POP ALL IEEE Xplore All Conference Proceedings IEEE Xplore IEEE Proceedings Order Plans (POP All) 1998-Present |
| DatabaseTitleList | |
| Database_xml | – sequence: 1 dbid: RIE name: IEEE Xplore Digital Library url: https://ieeexplore.ieee.org/ sourceTypes: Publisher |
| DeliveryMethod | fulltext_linktorsrc |
| EISBN | 9781728151250 1728151252 |
| EndPage | 251 |
| ExternalDocumentID | 8904862 |
| Genre | orig-research |
| GroupedDBID | 6IE 6IL ACM ALMA_UNASSIGNED_HOLDINGS APO CBEJK LHSKQ RIE RIL |
| ID | FETCH-LOGICAL-a283t-3199c8980e1e1192bbf0d542c91f190e7ef899c6ddf849b5fcdcb2d64fa2651f3 |
| IEDL.DBID | RIE |
| ISICitedReferencesCount | 10 |
| ISICitedReferencesURI | http://www.webofscience.com/api/gateway?GWVersion=2&SrcApp=Summon&SrcAuth=ProQuest&DestLinkType=CitingArticles&DestApp=WOS_CPL&KeyUT=000521634200029&url=https%3A%2F%2Fcvtisr.summon.serialssolutions.com%2F%23%21%2Fsearch%3Fho%3Df%26include.ft.matches%3Dt%26l%3Dnull%26q%3D |
| IngestDate | Wed Sep 03 07:09:58 EDT 2025 |
| IsPeerReviewed | false |
| IsScholarly | false |
| Language | English |
| LinkModel | DirectLink |
| MergedId | FETCHMERGED-LOGICAL-a283t-3199c8980e1e1192bbf0d542c91f190e7ef899c6ddf849b5fcdcb2d64fa2651f3 |
| PageCount | 9 |
| ParticipantIDs | ieee_primary_8904862 |
| PublicationCentury | 2000 |
| PublicationDate | 2019-Sep |
| PublicationDateYYYYMMDD | 2019-09-01 |
| PublicationDate_xml | – month: 09 year: 2019 text: 2019-Sep |
| PublicationDecade | 2010 |
| PublicationTitle | 2019 ACM/IEEE 22nd International Conference on Model Driven Engineering Languages and Systems Companion (MODELS-C) |
| PublicationTitleAbbrev | MODELS-C |
| PublicationYear | 2019 |
| Publisher | IEEE |
| Publisher_xml | – name: IEEE |
| SSID | ssib058463251 ssib045274562 |
| Score | 1.849624 |
| Snippet | We describe four scenario-based implementations of controllers for a player in a simplified RoboCup-type game. All four implementations are based on the... |
| SourceID | ieee |
| SourceType | Publisher |
| StartPage | 243 |
| SubjectTerms | Behavioral Programming BPjs Context modeling Context Oriented Modelling Couplings Deep reinforcement learning DRL Games Intelligent agents Manuals Model driven engineering Programming Robots Training |
| Title | Using Behavioural Programming with Solver, Context, and Deep Reinforcement Learning for Playing a Simplified RoboCup-Type Game |
| URI | https://ieeexplore.ieee.org/document/8904862 |
| WOSCitedRecordID | wos000521634200029&url=https%3A%2F%2Fcvtisr.summon.serialssolutions.com%2F%23%21%2Fsearch%3Fho%3Df%26include.ft.matches%3Dt%26l%3Dnull%26q%3D |
| hasFullText | 1 |
| inHoldings | 1 |
| isFullTextHit | |
| isPrint | |
| link | http://cvtisr.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwlV09T8MwELVKxcAEqEV8ywNjQ-PWSey5HzBAqVqQulW2c0aVaBKVFDHx2_G5oWVgYYtOGaK7yO_u_O4dITcsTgx0mA7AJDzgKDakrLGBcuDr0gMOUaz9solkNBKzmRzXSGs7CwMAnnwGt_jo7_LT3KyxVdYWEgXi3IG7lyTxZlbr59_hkSuvop30HeJq12F3JTTEQtl-fOoPHqZBD_lcKFIZ4oLwXytVPKIMD__3LUekuRvNo-Mt6ByTGmQN8uUv_mmldYhCGvgO8q6WaMdeK53myIFuUS9H9Vm2qMpS2gco6AS8eqrxjUJaCa6-Umei4zeFY1BU0ekCqefWJax0kuu8ty4CLGHpnVpCk7wMB8-9-6BarOBCILqlO3elNEKKEBgwl-JpbcM04h0jmXUJAiRgXRlm4jS1gksdWZMa3Uljbl0II2a7J6Se5RmcEio0Y34TnnIe55GSoKUNrTBScWu4OCMN9N282GhnzCu3nf9tviAHGJwNh-uS1MvVGq7IvvkoF--rax_wb5turQ8 |
| linkProvider | IEEE |
| linkToHtml | http://cvtisr.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwlV1NTwIxEG0ImuhJDRi_7cEjK1vosu2ZDzECEsCEG2m7U0MiLEEwnvztdsoKHrx420z2sJnZ9M1M37wh5I7VYgMVpgMwMQ84ig0pa2ygHPi69IBDVNN-2UTc64nxWPZzpLSdhQEATz6De3z0d_lJatbYKisLiQJx7sDdizivhJtprZ-_h0euwIp24neIrFWH3pnUEAtlufvcaHaGQR0ZXShTGeKK8F9LVTymtI7-9zXHpLgbzqP9LeyckBzMC-TLX_3TTO0QpTTwHWRezdCO3VY6TJEFXaJekOpzVaJqntAGwIIOwOunGt8qpJnk6it1Jtp_UzgIRRUdTpF8bl3KSgepTuvrRYBFLH1QMyiSl1ZzVG8H2WoFFwRRXbmTV0ojpAiBAXNJntY2TCJeMZJZlyJADNYVYqaWJFZwqSNrEqMrSY1bF8SI2eopyc_TOZwRKjRjfheech7nkZKgpQ2tMFJxa7g4JwX03WSxUc-YZG67-Nt8Sw7ao25n0nnsPV2SQwzUhtF1RfKr5Rquyb75WE3flzc--N_HHLBW |
| openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2019+ACM%2FIEEE+22nd+International+Conference+on+Model+Driven+Engineering+Languages+and+Systems+Companion+%28MODELS-C%29&rft.atitle=Using+Behavioural+Programming+with+Solver%2C+Context%2C+and+Deep+Reinforcement+Learning+for+Playing+a+Simplified+RoboCup-Type+Game&rft.au=Elyasaf%2C+Achiya&rft.au=Sadon%2C+Aviran&rft.au=Weiss%2C+Gera&rft.au=Yaacov%2C+Tom&rft.date=2019-09-01&rft.pub=IEEE&rft.spage=243&rft.epage=251&rft_id=info:doi/10.1109%2FMODELS-C.2019.00039&rft.externalDocID=8904862 |