[R] Validation / Training - test data

Sam Sam_Smith at me.com
Wed Sep 29 15:03:56 CEST 2010


Thanks for this,

I had used 

> validate(model0, method="boot",B=200)

To get a index.corrected Brier score, 

However i am also wanting to bootstrap the predicted probabilities  output from  predict(model1, type = "response") to get a idea of confidence, or am i best just using se.fit = TRUE and then calculating the 95%CI? Does what i want to do make sense?

Thanks


On 29 Sep 2010, at 13:38, Frank Harrell wrote:


Split sample validation is highly unstable with your sample size.

The rms package can help with bootstrapping or cross-validation, assuming
you have all modeling steps repreated for each resample.

Frank

-----
Frank Harrell
Department of Biostatistics, Vanderbilt University
-- 
View this message in context: http://r.789695.n4.nabble.com/Validation-Training-test-data-tp2718523p2718905.html
Sent from the R help mailing list archive at Nabble.com.

______________________________________________
R-help at r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.



More information about the R-help mailing list