Repository logo
  • Log In
    Log in via Symplectic to deposit your publication(s).
Repository logo
  • Communities & Collections
  • Research Outputs
  • Statistics
  • Log In
    Log in via Symplectic to deposit your publication(s).
  1. Home
  2. Faculty of Engineering
  3. Computing
  4. Computing
  5. Enablingmarkovian representations under imperfect information
 
  • Details
Enablingmarkovian representations under imperfect information
File(s)
EPOMGs_2021.pdf (170.99 KB)
Accepted version
Author(s)
Belardinelli, Francesco
Leon, Borja G
Malvone, Vadim
Type
Conference Paper
Abstract
Markovian systems are widely used in reinforcement learning (RL), when the successful completion of a task
depends exclusively on the last interaction between an autonomous agent and its environment. Unfortunately,
real-world instructions are typically complex and often better described as non-Markovian. In this paper we
present an extension method that allows solving partially-observable non-Markovian reward decision processes (PONMRDPs) by solving equivalent Markovian models. This potentially facilitates Markovian-based
state-of-the-art techniques, including RL, to find optimal behaviours for problems best described as PONMRDP. We provide formal optimality guarantees of our extension methods together with a counterexample
illustrating that naive extensions from existing techniques in fully-observable environments cannot provide
such guarantees.
Editor(s)
Rocha, AP
Steels, L
VandenHerik, J
Date Acceptance
2022-02-01
Citation
ICAART: PROCEEDINGS OF THE 14TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 2, pp.450-457
URI
http://hdl.handle.net/10044/1/97300
URL
https://www.scitepress.org/PublicationsDetail.aspx?ID=1lDLWpPvaP8=&t=1
DOI
https://www.dx.doi.org/10.5220/0010882200003116
ISSN
2184-433X
Publisher
SCITEPRESS
Start Page
450
End Page
457
Journal / Book Title
ICAART: PROCEEDINGS OF THE 14TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 2
Copyright Statement
© 2022 The Author(s). This work is published under CC BY-NC-ND 4.0 International licence.
License URL
https://creativecommons.org/licenses/by-nc-nd/4.0/
Identifier
http://gateway.webofknowledge.com/gateway/Gateway.cgi?GWVersion=2&SrcApp=PARTNER_APP&SrcAuth=LinksAMR&KeyUT=WOS:000774441800041&DestLinkType=FullRecord&DestApp=ALL_WOS&UsrCustomerID=1ba7043ffcc86c417c072aa74d649202
Source
14th International Conference on Agents and Artificial Intelligence (ICAART)
Subjects
Science & Technology
Technology
Computer Science, Artificial Intelligence
Computer Science, Interdisciplinary Applications
Computer Science, Software Engineering
Computer Science, Theory & Methods
Computer Science
Markov Decision Processes
Partial Observability
Extended Partially Observable Decision Process
non-Markovian Rewards
Publication Status
Published
Start Date
2022-02-03
Finish Date
2022-02-05
Coverage Spatial
ELECTR NETWORK
About
Spiral Depositing with Spiral Publishing with Spiral Symplectic
Contact us
Open access team Report an issue
Other Services
Scholarly Communications Library Services
logo

Imperial College London

South Kensington Campus

London SW7 2AZ, UK

tel: +44 (0)20 7589 5111

Accessibility Modern slavery statement Cookie Policy

Built with DSpace-CRIS software - Extension maintained and optimized by 4Science

  • Cookie settings
  • Privacy policy
  • End User Agreement
  • Send Feedback