## CLASS-L@LISTS.SUNYSB.EDU

 Options: Use Monospaced Font Show Text Part by Default Condense Mail Headers Message: [<< First] [< Prev] [Next >] [Last >>] Topic: [<< First] [< Prev] [Next >] [Last >>] Author: [<< First] [< Prev] [Next >] [Last >>]

```Many thanks for your kind replies. Yes, I mean "a priori" probabilities.
To make a decision, the Bayesian decision rule is

P(X|OBJECT)/P(X|NON-OBJECT)> lamda

Usually lamda=P(NON-OBJECT)/P(OBJECT). Here I just want to know how to
decide this lamda and the priori probabilities P(OBJECT) and
P(NON-OBJECT). For example in face detection or recognition, this can be
used to decide whether a new pattern is a face or non-face by computing
its likelihood ratios.From a training set the probabilistic model
parameters can be learnt. I am not sure whether the priori probabilities
should also be derived from the training set. For example, if in a
training set there are 3000 object patterns and 15000 non-object
patterns. In this case can we say the threshold, lamda, should be
15000/3000=5? Or can we still assume a equal priori which means lamda
should be 1?

Many Thanks.

On Wed, 2003-06-25 at 02:38, leo horseman wrote:
>
>
>
> Could you be a little more specific? Do you mean a priori
> probabilities?  Are you referring to the chance probabilities of an
> object belonging to a particular class or not belonging to that
> class?  What do you mean by a "training set"?  Can you give an example
> of a specific problem in terms of number of objects (entities, etc.)
> to be classified; number and type of variates  from which a pattern is
> derived; type of values assigned to the variates, etc.
>
>                               M. Childress
```