Interobserver agreement in behavioral research: Importance and calculation. Watkins, M. W. & Pacheco, M. Journal of Behavioral Education, 10(4):205–212, 2000.
Interobserver agreement in behavioral research: Importance and calculation [link]Paper  doi  abstract   bibtex   
Behavioral researchers have developed a sophisticated methodology to evaluate behavioral change which is dependent upon accurate measurement of behavior. Direct observation of behavior has traditionally been the mainstay of behavioral measurement. Consequently, researchers must attend to the psychometric properties, such as interobserver agreement, of observational measures to ensure reliable and valid measurement. Of the many indices of interobserver agreement, percentage of agreement is the most popular. Its use persists despite repeated admonitions and empirical evidence indicating that it is not the most psychometrically sound statistic to determine interobserver agreement due to its inability to take chance into account. Cohen's (1960) kappa has long been proposed as the more psychometrically sound statistic for assessing interobserver agreement. Kappa is described and computational methods are presented.
@article{watkins_interobserver_2000,
	title = {Interobserver agreement in behavioral research: {Importance} and calculation},
	volume = {10},
	issn = {1053-0819, 1573-3513},
	shorttitle = {Interobserver {Agreement} in {Behavioral} {Research}},
	url = {http://link.springer.com/article/10.1023/A%3A1012295615144},
	doi = {10.1023/A:1012295615144},
	abstract = {Behavioral researchers have developed a sophisticated methodology to evaluate behavioral change which is dependent upon accurate measurement of behavior. Direct observation of behavior has traditionally been the mainstay of behavioral measurement. Consequently, researchers must attend to the psychometric properties, such as interobserver agreement, of observational measures to ensure reliable and valid measurement. Of the many indices of interobserver agreement, percentage of agreement is the most popular. Its use persists despite repeated admonitions and empirical evidence indicating that it is not the most psychometrically sound statistic to determine interobserver agreement due to its inability to take chance into account. Cohen's (1960) kappa has long been proposed as the more psychometrically sound statistic for assessing interobserver agreement. Kappa is described and computational methods are presented.},
	language = {en},
	number = {4},
	urldate = {2014-10-26},
	journal = {Journal of Behavioral Education},
	author = {Watkins, Marley W. and Pacheco, Miriam},
	year = {2000},
	keywords = {Clinical Psychology, Health Psychology, Kappa, Pedagogic Psychology, Psychology of Personality, interobserver agreement, interrater reliability, observer agreement, percent agreement},
	pages = {205--212},
}

Downloads: 0