Synthesis from Satisficing and Temporal Goals

1Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.

Abstract

Reactive synthesis from high-level specifications that combine hard constraints expressed in Linear Temporal Logic (LTL) with soft constraints expressed by discounted-sum (DS) rewards has applications in planning and reinforcement learning. An existing approach combines techniques from LTL synthesis with optimization for the DS rewards but has failed to yield a sound algorithm. An alternative approach combining LTL synthesis with satisficing DS rewards (rewards that achieve a threshold) is sound and complete for integer discount factors, but, in practice, a fractional discount factor is desired. This work extends the existing satisficing approach, presenting the first sound algorithm for synthesis from LTL and DS rewards with fractional discount factors. The utility of our algorithm is demonstrated on robotic planning domains.

Cite

CITATION STYLE

APA

Bansal, S., Kavraki, L., Vardi, M. Y., & Wells, A. (2022). Synthesis from Satisficing and Temporal Goals. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022 (Vol. 36, pp. 9679–9686). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v36i9.21202

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free