coalition is defined as the coalition value minus the
execution cost of all the coalition’s members. When
an agent of the coalition does not know with certainty
the execution costs of the other members, it is uncer-
tain regarding both the coalition’s net benefits and its
net benefits. A protocol allowing agents to negoti-
ate and form coalition in such a case has been pro-
posed in (Kraus et al., 2003) and (Kraus et al., 2004).
Another source of uncertainty on coalition value can
be the imperfect or deceiving information. A study
for this case has been proposed in (Blankenburg and
Klusch, 2004). In (Chalkiadakis and Boutilier, 2004),
authors proposed a reinforcement learning model to
allow agents to refine their beliefs about others’ capa-
bilities.
Although these previous works deal with an impor-
tant uncertainty issue (uncertain coalition value), they
have several restrictive assumptions regarding another
possible sources of uncertainty as the uncertain re-
sources consumption (uncertain task execution) that
can be due to the uncertain agent’s behavior and to
the environment’s dynamism. In addition, they do not
take into account the effects of forming a coalition on
the future possible formations, a long-term coalition
formation planning can not then be provided. In ap-
plications as planetary rovers, for example, an agent is
confronted with an ambiguous environment where he
can not control his resources consumption when exe-
cuting tasks as good as he does in laboratory. A coali-
tion formation planning is important so that agents
adapt coalition formation to their uncertain behaviors.
The problem is more complex when resources con-
sumption is uncertain for all the agents. Unfor-
tunately, in such a system, an agent can’t be sure
whether he (or another agent) will be able to execute
all the subtasks that are allocated to him or he will
ignore some of them. So, forming coalitions to max-
imize the agents’ real reward is a complex (even un-
realizable) operation. In fact, a task is considered as
non executed if at least one of its subtasks is not ex-
ecuted. That is why, forming a coalition to execute
a task is a necessary but not sufficient constraint to
obtain a reward, and the agents’ reward must be sub-
jected to the task execution and not only to the coali-
tion formation and task allocation. In this paper, we
take into account these issues and we present a prob-
abilistic model, based on Markov Decision Process
(MDP), that provides a coalition formation planning
for environments where resources consumption is un-
certain. We will show that according to each possible
resources consumption, agents can decide by an opti-
mal way which coalition they must form.
We begin in Section 2 with a presentation of our
framework. In section 3, we sketch our solution ap-
proach. We explain how to form coalition via MDP
in Section 4.
2 FRAMEWORK
We consider a situation where a set of m fully-
cooperative agents, A = {a
1
,...,a
m
} have to coop-
erate to execute a finite set of tasks T = {T
1
,...,T
n
}
in an uncertain environment. The tasks will be
allocated in a commonly known order: without
loss of generality, we assume that this ordering is
T
1
,T
2
, ··· ,T
n
. Each agent a
k
has a bounded quan-
tity of resources R
k
that he uses to execute tasks.
Each task consists of subtasks: for simplicity, we
assume that every task T
i
∈ T is composed by q
subtasks such as T
i
= {t
1
i
,...,t
q
i
}. Agent a
k
is
able to perform only a subset E
k
i
⊂ T
i
of the sub-
tasks of a given task T
i
. We assume that each task
T
i
∈ T satisfies the condition T
i
⊆∪
a
k
∈A
E
k
i
, oth-
erwise it is an unrealizable task. For each subtask
t
l
i
,T
i
, ∈ T,l =1,...,q, we can define the set of
agents, AE(t
l
i
), that are able to perform t
l
i
as follows:
AE(t
l
i
)={a
k
∈ A|t
l
i
∈ E
k
i
}. Since an agent can’t
execute a task T
i
by himself, a coalition of agents
must be formed in order to execute this task. Such
a coalition can be defined as a q−tuple: a
1
,...,a
q
where agent a
l
∈ A executes subtask t
l
i
∈ E
a
l
i
.We
let C(T
i
) denote the set of all possible coalitions that
can perform task T
i
, it can be defined as follows:
C(T
i
)={a
1
,...,a
q
|a
l
∈ A, t
l
i
∈ T
i
,t
l
i
∈ E
a
l
i
,l =
1,...,q}. A task is considered as realized if and only
if all its subtasks have been performed. For each re-
alized task T
i
, agents obtain a reward. We consider
a general situation where the tasks can be executed
with different qualities. For example, two agents can
take photos for the same object, but the resolution can
be different. The reward corresponding to the execu-
tion of a task depends then on the coalition that ex-
ecutes the task. We assume that agents have a func-
tion w(T
i
,c) that expresses the reward that can be ob-
tained if the coalition c executes task T
i
.
3 SOLUTION APPROACH
The key idea, in our approach, is to view the forma-
tion of a coalition to execute a task as a decision to
make that provides an expected reward instead of a
real gain. What one expects to gain by forming col-
lation c to execute task T
i
? In fact, when T
i
is allo-
cated to c, the agents expect to obtain two values. The
first one is the value w(T
i
,c) which is subjected to
the execution of task T
i
. The second expected value
expresses the gain that can be obtained from future
formation and allocation taking into consideration re-
sources quantity consumed to execute T
i
. Indeed,
when a coalition executes a task, the agents’ available
resources is reduced. The chances to execute another
COALITION FORMATION WITH UNCERTAIN TASK EXECUTION
165