Many thanks for your kind replies. Yes, I mean "a priori" probabilities.
To make a decision, the Bayesian decision rule is


Usually lamda=P(NON-OBJECT)/P(OBJECT). Here I just want to know how to
decide this lamda and the priori probabilities P(OBJECT) and
P(NON-OBJECT). For example in face detection or recognition, this can be
used to decide whether a new pattern is a face or non-face by computing
its likelihood ratios.From a training set the probabilistic model
parameters can be learnt. I am not sure whether the priori probabilities
should also be derived from the training set. For example, if in a
training set there are 3000 object patterns and 15000 non-object
patterns. In this case can we say the threshold, lamda, should be
15000/3000=5? Or can we still assume a equal priori which means lamda
should be 1?

Many Thanks.

On Wed, 2003-06-25 at 02:38, leo horseman wrote:
> Could you be a little more specific? Do you mean a priori
> probabilities?  Are you referring to the chance probabilities of an
> object belonging to a particular class or not belonging to that
> class?  What do you mean by a "training set"?  Can you give an example
> of a specific problem in terms of number of objects (entities, etc.)
> to be classified; number and type of variates  from which a pattern is
> derived; type of values assigned to the variates, etc.
>                               M. Childress