Using Qualitative Preferences to Guide Schedule Optimization

As robots and computers become more capable, we see more mixed human-robot- computer teams. One strategy for coordinating mixed teams is standardizing priorities across agents. However, the encoding of priorities may lead to discrepancies in interpretation across different types of agents. For examp...

Full description

Bibliographic Details
Main Author: Wells, Tesla
Other Authors: Williams, Brian C.
Format: Thesis
Published: Massachusetts Institute of Technology 2023
Online Access:https://hdl.handle.net/1721.1/150116
Description
Summary:As robots and computers become more capable, we see more mixed human-robot- computer teams. One strategy for coordinating mixed teams is standardizing priorities across agents. However, the encoding of priorities may lead to discrepancies in interpretation across different types of agents. For example, a machine may have difficulty generating a plan that incorporates the subtleties of a natural language description. A human, given a purely quantitative encoding, may have difficulty intuiting trends or generating a plan without a computation aid. These discrepancies in priority interpretation may lead to a plan that only one class of agent deems optimal. In this case of distributed planning, differing interpretations may lead to significant differences in the plans produced by different classes of agents. What’s more, low fidelity interpretation of priorities can prohibit model correction or refinement. A robot may not recognize its learnt-physics-model conflicts with underlying assumptions in a natural language description of priorities. For humans, it is difficult to identify incorrect or unrefined priority models from the same streams of numerical relations computers find useful. Most strategies currently in use for bridging this gap involve machines learning human preferences from large data sets or require labor-intensive custom utility encodings to be written, explained, and revised by a trained expert. The former often homogenizes models of human preferences and fails to incorporate corrections accurately, efficiently, and in context. The latter prohibits the usage of robots or computer aids in casual or dynamic settings without the presence/supervision of a human trained in working with the system. In both settings, this inhibits average humans from having personalized, human-computer or human-robot interactions. In this thesis, we attempt to improve human-robot interactions by encoding utility as a series of qualitative, Ceteris Paribus preference statements over a design space. We posit this formalism is both readily understood by human agents and easily reasoned over by machines. Previous work establishes the ability to compute machine readable utility functions from said statements by efficiently generating topological orderings. We detail our implementation of a machine “agent” who uses said utility function to compute optimal schedules to Conditional Temporal Problem problems for a mixed-agent team. We then build off of the procedure for generating utility functions to generate admissible heuristics that increase performance. We show this encoding enables us to explain which human preferences differentiate feasibly scheduled, high-utility plans. We present a suit of algorithms capable of explaining model behavior according to five different relevance standards. Additionally, we construct and algorithm for identifying where additional preference specification would resolve assumptions used in underspecified areas of the model. These explanations not only improve human understanding, but also facilitate identification of inaccuracies in the machine’s utility model. We build off our explanations to presents options for model-repair. We show preference-addition and preference-relaxation informed by explanation results in specific, targeted plan changes.