Synthesis of Provably Correct Autonomy Protocols for Shared Control. (arXiv:1905.06471v1 [cs.RO])

We synthesize shared control protocols subject to probabilistic temporal
logic specifications. More specifically, we develop a framework in which a
human and an autonomy protocol can issue commands to carry out a certain task.
We blend these commands into a joint input to a robot. We model the interaction
between the human and the robot as a Markov decision process (MDP) that
represents the shared control scenario. Using inverse reinforcement learning,
we obtain an abstraction of the human’s behavior and decisions. We use
randomized strategies to account for randomness in human’s decisions, caused by
factors such as complexity of the task specifications or imperfect interfaces.
We design the autonomy protocol to ensure that the resulting robot behavior
satisfies given safety and performance specifications in probabilistic temporal
logic. Additionally, the resulting strategies generate behavior as similar to
the behavior induced by the human’s commands as possible. We solve the
underlying problem efficiently using quasiconvex programming. Case studies
involving autonomous wheelchair navigation and unmanned aerial vehicle mission
planning showcase the applicability of our approach.

Source link

Related posts

Can Artificial Intelligence Make Screening Faster, More Accurate, and More Accessible?


Measuring ability of artificial intelligence to learn is difficult


Rigorous prognostication and modeling of gas adsorption on activated carbon and Zeolite-5A.


This website uses cookies to improve your experience. We'll assume you're ok with this, but you can opt-out if you wish. Accept Read More

Privacy & Cookies Policy