Loading...
Please wait, while we are loading the content...
Similar Documents
Learning Factored Markov Decision Processes with Unawareness Extended Abstract
| Content Provider | Semantic Scholar |
|---|---|
| Author | Innes, Craig Lascarides, Alex |
| Copyright Year | 2019 |
| Abstract | Methods for learning and planning in sequential decision problems often assume the learner is fully aware of all possible states and actions in advance. This assumption is sometimes untenable: evidence gathered via domain exploration or external advice may reveal not just information about which of the currently known states are probable, but that entirely new states or actions are possible. This paper provides a model-based method for learning factored markov decision problems from both domain exploration and contextually relevant expert corrections in a way which guarantees convergence to near-optimal behaviour, even when the agent is initially unaware of actions or belief variables that are critical to achieving success. Our experiments show that our agent converges quickly on the optimal policy for both large and small decision problems. We also explore how an expert's tolerance towards the agent's mistakes affects the agent's ability to achieve optimal behaviour.1 |
| File Format | PDF HTM / HTML |
| Alternate Webpage(s) | http://homepages.inf.ed.ac.uk/alex/papers/aamas_fmdp.pdf |
| Alternate Webpage(s) | http://www.ifaamas.org/Proceedings/aamas2019/pdfs/p2030.pdf |
| Language | English |
| Access Restriction | Open |
| Content Type | Text |
| Resource Type | Article |