Contingency Analysis in the Design-to-Criteria Scheduler
We define schedule robustness as a characteristic of a schedule in which the schedule allows for recovery from execution failure of one of the scheduled actions. In evaluating a schedule, we want to take into account whether there exist alternative ways of completing the schedule, i.e., achieving the high level task, if the schedule should fail during the course of execution. This type of analysis, called contingency planning can be expensive because it could involve an exhaustive search for the appropriate method that would improve schedule robustness without diminishing the criteria requirements [Bresina94]. However, the technique we describe here implements an algorithm which eliminates the need to do an exhaustive search, even though it is more expensive than our non-contingency scheduling approach.
We found that our work done using the Design-to-Criteria framework addresses similar questions namely
For the example described in the earlier section, we will focus on maximizing quality within a hard deadline of 18 minutes. The two possible schedules are {A1,A2,A3} and {B}. The ELB takes into account the various possible permutations of method outcomes along with the quality achieved. Figure 2 describes the computation of the ELB ratings for the schedule {A1,A2,A3}.
Consider the first entry of the table. It handles the case when method A1 achieves a quality of 2, which occurs with a probability of 0.5 as described in the TÆMS task structure. Method A2 achieves a quality of 0 with probability 0.25. The probability of the methods achieving these qualities simultaneously in a single execution is 0.125, given in column 4. The expected quality of the schedule {A1,A2,A3} is 0 in this case, described in column 5. The duration and cost distributions and their expected values are computed in a similar fashion. The ELB ratings for schedules {A1,A2,A3} and {B} are given below.
The schedule {A1,A2,A3} is chosen and executed since it has the
best expected lower bound rating of 0.97. A1 executes successfully , then
A2 executes and suppose A2 fails (i.e. it results in 0 quality), which
happens 25% of the time. Then A3 fails to get enabled and the schedule
breaks since there is no time left to reschedule {B} as an alternate schedule.
Because of the one-pass low-order polynomial method sequencing approach used by the scheduler to control scheduling combinatorics, the standard Design-to-Criteria scheduler will only produce one permutation of the methods A1, A2, and A3. However, if the scheduler did produce multiple permutations, the schedules {A1,A2,A3} and {A2,A1,A3} would receive the same expected lower bound value. Hence the contention is that there is no difference in performance if either of the two was chosen, or produced by the method ordering heuristics. However on more detailed evaluation of the schedules, we see that {A2,A1,A3} allows for recovery and contingency scheduling which schedule {A1,A2,A3} does not permit (Figure 3) for the given deadline. If {A2,A1,A3} is the schedule being executed and A2 fails, there is time to schedule method {B} and complete task TG1. This clearly implies that schedule {A2,A1,A3} should have a better expected performance rating than {A1,A2,A3} as the schedule {A2,A1,A3} includes the recovery option from failure in its structure.
Quality : (32% 0.5)(23% 1.0)(45% 2.0)
Duration: (100% 18)
The tree structure in Figure 3 presents all the options of schedule generation that will meet the criteria of a duration limit of 18 minutes. From this diagram, we see that schedule {A1,A2,A3} does not have an option to reschedule and still meet the deadline, if method A2 produces an undesirable outcome.
So we consider a simple reordering of schedule {A1,A2,A3} which is {A2,A1,A3}. To assess the effects of rescheduling when A2 fails on this schedule {A2,A1,A3}, we combine the ratings for schedules and based on their likelihoods of occurrence. So a schedule starting with A2 gets a rating of We use a similar analysis to get the values of schedules starting with A1 and B =1*0.60=0.60
This type of evaluation of the schedule is what we call the Approximate Expected Bound(AEB), which will be formally defined in the next section.
So schedule {A2,A1,A3} has a better expected performance than {A1,A2,A3}. The ELB computation of the Design-to-Criteria scheduler evaluates the performance measure of both {A1,A2,A3} and {A2,A1,A3} to be the same as it does not take into account the recovery options present within {A2,A1,A3} while evaluating it. This leads us to believe that the ELB perhaps is not the most appropriate performance measure for all task structures, particularly where hard deadlines or cost limits (in contrast to soft preferences) are important.
The bar graph on the left of Figure 4 shows the statistical performance of the schedule with the highest ELB for 100 simulation runs, namely {A1,A2,A3}. A simulation run is a simulated execution of the schedule with highest ELB and the actual quality, cost and durations values are averaged over the number of simulations to obtain a statistical rating. We note that the schedule fails to achieve any quality about 20% of the time. The mean quality achieved by using this performance measure is 0.98.
The bar graph on the right Figure 4 describes the statistical performance data for the schedule with the highest AEB over 100 simulation runs, namely {A2,A1,A3}. Here 100 simulated executions of the schedule with the highest AEB rating produce the data. As seen in the histogram, the quality of the schedule with the best AEB rating is always a non-zero value due to the built-in contingency and the mean quality achieved here is 1.96.
Currently we are investigating the use of contingency analysis to build
better schedules. This work is
complimentary to the work presented here. Consider a situation where
a highly uncertain action is embedded in
the middle of a schedule. Because of complex task interactions and
the complex semantics of the functions that determine how quality is accumulated
by tasks via their subtasks, determining the criticality of this highly
uncertain action is neither trivial nor computationally cheap. In some
cases, a failure may negate any useful
work done to that point. In situations where hard deadlines are present,
failure at that point may translate into
overall problem solving failure if the problem solver lacks sufficient
time remaining to try a different solution
path. In hard deadline situations, it may be worthwhile to understand
some of the downstream ramifications of a failure at that point and to
attempt resequence the schedule accordingly - perhaps resulting in a lower
quality
schedule that has more possible contingent schedules in the case of
action failure. Without explicit
representation of uncertainty, this type of analysis would not be possible.
Another area of future uncertainty-related work in Design-to-Criteria
scheduling involves leveraging the
uncertainty-chanced TÆMS models in multi-agent scheduling. In
multi-agent systems the scheduler is
typically coupled with a multi-agent coordination module that forms
commitments to perform work with other
agents; local concerns are thus modulated by non-local problem solving.
Uncertainty in this context could be
used to reason about the utility of the commitments made with other
agents and to understand how the
uncertainty about commitments made by other agents affects local problem
solving.
Other, more general, future efforts in Design-to-Criteria will center
around negotiation between the scheduler
and its clients, which may be other AI problem solvers or humans. Negotiation
during the scheduling process
can iteratively refine client goal criteria based on what is actually
being produced by the scheduler. This is
important because often if the scheduler cannot produce schedules that
satisfice well enough with respect to the goal criteria, due to task limitations
or resource constraints, the client may prefer to submit a different set
of goal criteria and try again, exploring the solution space prior to selecting
a course of action.
Mostafa, Hala; Lesser, Victor. Approximately Solving Sequential Games with Incomplete Information. Proceedings of the AAMAS’08 Workshop on Multi-Agent Sequential Decision Making in Uncertain Multi-Agent Domains, Shen, J.; Varakantham, P.; Maheswaran, R., ed., IFAAMAS, pp. 92-106. 2008.
Wagner, Thomas; Raja, Anita; Lesser, Victor. Modeling Uncertainty and its Implications to Sophisticated Control in TAEMS Agents. Autonomous Agents and Multi-Agent Systems, Volume 13, Number 3, Springer: Netherlands, pp. 235-292. 2006. On-line version available April 2006.
Raja, A.; Wagner, T.; Lesser, V. Reasoning about Uncertainty in Agent Control. Proceedings of the 5th International Conference on Information Systems, Analysis, and Synthesis, Computer Science and Engineering: Part 1, Volume VII, pp. 156-161. 2001.
Raja, Anita, Lesser, Victor, and Wagner, Thomas. Toward Robust Agent Control in Open Environments. Proceedings of 5th International Conference of Autonomous Agents(AA 2000). Also Umass CS Technical Report 1999-059, pp. 84-91. 2000.
Raja, Anita, Wagner, Thomas and Lesser, Victor. Reasoning about Uncertainty in Design-to-Criteria Scheduling. Proceedings of AAAI 2000 Spring Symposium on Real-Time Autonomous Systems, AAAI, pp. 76-83. 2000.
Wagner, Thomas; Raja, Anita; and Lesser, Victor. Modeling Uncertainty and its Implications to Design-to-Criteria Scheduling. UMASS Technical Report TR 1998-51, Number TR 1998-51, University of Massachusetts. 1999.
Raja, Anita, Wagner, Thomas, Lesser, Victor. Ensuring Robust Agent Control in Uncertain Environments.. University of Massachusetts Computer Science TR-99-027. 1999.
Xuan, Ping, Lesser, Victor. Incorporating Uncertainty in Agent Commitments. Intelligent Agents VI --- Proceedings of the Sixth International Workshop on Agent Theories, Architectures, and Languages (ATAL-99), Lecture Notes in Artificial Intelligence, N. R. Jennings and Y. Lesperance (eds.), Springer-Verlag, Berlin, pp. 57-70. 1999.