Identifying Feature Relevance Using a Random Forest. Rogers, J. and Gunn, S. Volume 3940, Saunders, C.; Grobelnik, M.; Gunn, S.; and Shawe-Taylor, J., editors. Subspace, Latent Structure and Feature Selection, pages 173--184. Springer, Berlin / Heidelberg, 2006.
doi  abstract   bibtex   
It is known that feature selection and feature relevance can benefit the performance and interpretation of machine learning algorithms. Here we consider feature selection within a Random Forest framework. A feature selection technique is introduced that combines hypothesis testing with an approximation to the expected performance of an irrelevant feature during Random Forest construction. It is demonstrated that the lack of implicit feature selection within Random Forest has an adverse effect on the accuracy and efficiency of the algorithm. It is also shown that irrelevant features can slow the rate of error convergence and a theoretical justification of this effect is given.
@inbook{Rogers:2006aa,
	Abstract = {It is known that feature selection and feature relevance can benefit the performance and interpretation of machine learning algorithms. Here we consider feature selection within a Random Forest framework. A feature selection technique is introduced that combines hypothesis testing with an approximation to the expected performance of an irrelevant feature during Random Forest construction.

It is demonstrated that the lack of implicit feature selection within Random Forest has an adverse effect on the accuracy and efficiency of the algorithm. It is also shown that irrelevant features can slow the rate of error convergence and a theoretical justification of this effect is given.},
	Address = {Berlin / Heidelberg},
	Author = {Rogers, Jeremy and Gunn, Steve},
	Chapter = {Identifying Feature Relevance Using a Random Forest},
	Date-Added = {2008-05-10 13:39:38 -0400},
	Date-Modified = {2009-04-16 15:14:55 -0400},
	Doi = {10.1007/11752790_12},
	Editor = {Saunders, C. and Grobelnik, M. and Gunn, S. and Shawe-Taylor, J.},
	Keywords = {feature selection; descriptor importance; variable importance; bagging},
	Pages = {173--184},
	Publisher = {Springer},
	Series = {Lecture Notes in Computer Science},
	Timescited = {0},
	Title = {Subspace, Latent Structure and Feature Selection},
	Volume = {3940},
	Year = {2006},
	Bdsk-File-1 = {YnBsaXN0MDDUAQIDBAUIJidUJHRvcFgkb2JqZWN0c1gkdmVyc2lvblkkYXJjaGl2ZXLRBgdUcm9vdIABqAkKFRYXGyIjVSRudWxs0wsMDQ4RElpOUy5vYmplY3RzViRjbGFzc1dOUy5rZXlzog8QgASABoAHohMUgAKAA1lhbGlhc0RhdGFccmVsYXRpdmVQYXRo0hgMGRpXTlMuZGF0YU8RAZwAAAAAAZwAAgAAA212IAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAMU5bQNIKwAAABCNbRgxMC4xMDA3LTExNzUyNzkwXzEyMC5wZGYAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAEI15xEtXulBERiBwcnZ3AAIAAwAACSAAAAAAAAAAAAAAAAAAAAAIYXJ0aWNsZXMAEAAIAADFObNTAAAAEQAIAADES4/6AAAAAQAQABCNbQAKTIAACkxpAAB8EwACADttdiA6VXNlcnM6cmd1aGE6RG9jdW1lbnRzOmFydGljbGVzOjEwLjEwMDctMTE3NTI3OTBfMTIwLnBkZgAADgAyABgAMQAwAC4AMQAwADAANwAtADEAMQA3ADUAMgA3ADkAMABfADEAMgAwAC4AcABkAGYADwAIAAMAbQB2ACAAEgA3VXNlcnMvcmd1aGEvRG9jdW1lbnRzL2FydGljbGVzLzEwLjEwMDctMTE3NTI3OTBfMTIwLnBkZgAAEwABLwAAFQACAAz//wAAgAXSHB0eH1gkY2xhc3Nlc1okY2xhc3NuYW1lox8gIV1OU011dGFibGVEYXRhVk5TRGF0YVhOU09iamVjdF8QMS4uLy4uL0RvY3VtZW50cy9hcnRpY2xlcy8xMC4xMDA3LTExNzUyNzkwXzEyMC5wZGbSHB0kJaIlIVxOU0RpY3Rpb25hcnkSAAGGoF8QD05TS2V5ZWRBcmNoaXZlcgAIABEAFgAfACgAMgA1ADoAPABFAEsAUgBdAGQAbABvAHEAcwB1AHgAegB8AIYAkwCYAKACQAJCAkcCUAJbAl8CbQJ0An0CsQK2ArkCxgLLAAAAAAAAAgEAAAAAAAAAKAAAAAAAAAAAAAAAAAAAAt0=}}
Downloads: 0