[R] Best performance measure?

Frank E Harrell Jr f.harrell at vanderbilt.edu
Wed Aug 19 21:11:51 CEST 2009


Noah Silverman wrote:
> Frank,
> 
> That makes sense.
> 
> I just had a look at the actual algorithm calculating the Briar score. 
> 
> One thing that confuses me is how the score is calculated.
> 
> 
> 
> If I understand the code correctly, it is just:  sum((p - y)^2)/n
> 
> If I have an example with a label of 1 and a probability prediction of 
> .4, it is (.4 - 1)^2 
> (I know it is the average of these value across all the examples)

Yes and I seem to remember the original score is 1 minus that.

> 
> Wouldn't it make more sense to stratify the probabilities and then check 
> the accuracy of each level.

The stratification will bring a great deal of noise into the problem. 
Better: loess calibration curves or decomposition of the Brier score 
into discrimination and calibration components (which is not in the 
software).

Frank

> 
> i.e.  For predicted probabilities of .10 to .20 the data was actually 
> labeled true .18 percent of the time. mean(label)
> 
> 
> 
> 
> 
> 
> On 8/19/09 11:51 AM, Frank E Harrell Jr wrote:
>> Noah Silverman wrote:
>>> Thanks for the suggestion.
>>>
>>> You explained that Briar combines both accuracy and discrimination 
>>> ability.  If I understand you right, that is in relation to binary 
>>> classification.
>>>
>>> I'm not concerned with binary classification, but the accuracy of the 
>>> probability predictions.
>>>
>>> Is there some kind of score that measures just the accuracy?
>>>
>>> Thanks!
>>>
>>> -N
>>
>> The Brier score has nothing to do with classification.  It is a 
>> probability accuracy score.
>>
>> Frank
>>
>>>
>>> On 8/19/09 10:42 AM, Frank E Harrell Jr wrote:
>>>> Noah Silverman wrote:
>>>>> Hello,
>>>>>
>>>>> I working on a model to predict probabilities.
>>>>>
>>>>> I don't really care about binary prediction accuracy.
>>>>>
>>>>> I do really care about the accuracy of my probability predictions.
>>>>>
>>>>> Frank was nice enough to point me to the val.prob function from the 
>>>>> Design library.  It looks very promising for my needs.
>>>>>
>>>>> I've put together some tests and run the val.prob analysis.  It 
>>>>> produces some very informative graphs along with a bunch of 
>>>>> performance measures.
>>>>>
>>>>> Unfortunately, I'm not sure which measure, if any, is the "best" 
>>>>> one.  I'm comparing hundreds of different models/parameter 
>>>>> combinations/etc.  So Ideally I'd like a single value or two as the 
>>>>> "performance measure" for each one.  That way I can pick the 
>>>>> "best"  model from all my experiments.
>>>>>
>>>>> As mentioned above, I'm mainly interested in the accuracy of my 
>>>>> probability predictions.
>>>>>
>>>>> Does anyone have an opinion about which measure I should look at??
>>>>> (I see Dxy, C, R2, D, U, Briar, Emax, Eavg, etc.)
>>>>>
>>>>> Thanks!!
>>>>>
>>>>> -N
>>>>
>>>> It all depends on the goal, i.e., the relative value you place on 
>>>> absolute accuracy vs. discrimination ability. The Brier score 
>>>> combines both and other than interpretability has many advantages.
>>>>
>>>> Frank
>>>>
>>>>>
>>>>> ______________________________________________
>>>>> R-help at r-project.org mailing list
>>>>> https://stat.ethz.ch/mailman/listinfo/r-help
>>>>> PLEASE do read the posting guide 
>>>>> http://www.R-project.org/posting-guide.html
>>>>> and provide commented, minimal, self-contained, reproducible code.
>>>>>
>>>>
>>>>
>>
>>


-- 
Frank E Harrell Jr   Professor and Chair           School of Medicine
                      Department of Biostatistics   Vanderbilt University




More information about the R-help mailing list