Assessing Readability of Medical Documents: A Ranking Approach. Zheng, J. & Yu, H The Journal of Medical Internet Research Medical Informatics, March, 2018. doi abstract bibtex BACKGROUND: The use of electronic health record (EHR) systems with patient engagement capabilities, including viewing, downloading, and transmitting health information, has recently grown tremendously. However, using these resources to engage patients in managing their own health remains challenging due to the complex and technical nature of the EHR narratives. OBJECTIVE: Our objective was to develop a machine learning-based system to assess readability levels of complex documents such as EHR notes. METHODS: We collected difficulty ratings of EHR notes and Wikipedia articles using crowdsourcing from 90 readers. We built a supervised model to assess readability based on relative orders of text difficulty using both surface text features and word embeddings. We evaluated system performance using the Kendall coefficient of concordance against human ratings. RESULTS: Our system achieved significantly higher concordance (.734) with human annotators than did a baseline using the Flesch-Kincaid Grade Level, a widely adopted readability formula (.531). The improvement was also consistent across different disease topics. This method's concordance with an individual human user's ratings was also higher than the concordance between different human annotators (.658). CONCLUSIONS: We explored methods to automatically assess the readability levels of clinical narratives. Our ranking-based system using simple textual features and easy-to-learn word embeddings outperformed a widely used readability formula. Our ranking-based method can predict relative difficulties of medical documents. It is not constrained to a predefined set of readability levels, a common design in many machine learning-based systems. Furthermore, the feature set does not rely on complex processing of the documents. One potential application of our readability ranking is personalization, allowing patients to better accommodate their own background knowledge.
@article{zheng_assessing_2018,
title = {Assessing {Readability} of {Medical} {Documents}: {A} {Ranking} {Approach}.},
doi = {DOI: 10.2196/medinform.8611},
abstract = {BACKGROUND:
The use of electronic health record (EHR) systems with patient engagement capabilities, including viewing, downloading, and transmitting health information, has recently grown tremendously. However, using these resources to engage patients in managing their own health remains challenging due to the complex and technical nature of the EHR narratives.
OBJECTIVE:
Our objective was to develop a machine learning-based system to assess readability levels of complex documents such as EHR notes.
METHODS:
We collected difficulty ratings of EHR notes and Wikipedia articles using crowdsourcing from 90 readers. We built a supervised model to assess readability based on relative orders of text difficulty using both surface text features and word embeddings. We evaluated system performance using the Kendall coefficient of concordance against human ratings.
RESULTS:
Our system achieved significantly higher concordance (.734) with human annotators than did a baseline using the Flesch-Kincaid Grade Level, a widely adopted readability formula (.531). The improvement was also consistent across different disease topics. This method's concordance with an individual human user's ratings was also higher than the concordance between different human annotators (.658).
CONCLUSIONS:
We explored methods to automatically assess the readability levels of clinical narratives. Our ranking-based system using simple textual features and easy-to-learn word embeddings outperformed a widely used readability formula. Our ranking-based method can predict relative difficulties of medical documents. It is not constrained to a predefined set of readability levels, a common design in many machine learning-based systems. Furthermore, the feature set does not rely on complex processing of the documents. One potential application of our readability ranking is personalization, allowing patients to better accommodate their own background knowledge.},
journal = {The Journal of Medical Internet Research Medical Informatics},
author = {Zheng, JP and Yu, H},
month = mar,
year = {2018},
pmid = {29572199 PMCID: PMC5889493},
}
Downloads: 0
{"_id":"NdTrvdrBDrLtiWtoA","bibbaseid":"zheng-yu-assessingreadabilityofmedicaldocumentsarankingapproach-2018","author_short":["Zheng, J.","Yu, H"],"bibdata":{"bibtype":"article","type":"article","title":"Assessing Readability of Medical Documents: A Ranking Approach.","doi":"DOI: 10.2196/medinform.8611","abstract":"BACKGROUND: The use of electronic health record (EHR) systems with patient engagement capabilities, including viewing, downloading, and transmitting health information, has recently grown tremendously. However, using these resources to engage patients in managing their own health remains challenging due to the complex and technical nature of the EHR narratives. OBJECTIVE: Our objective was to develop a machine learning-based system to assess readability levels of complex documents such as EHR notes. METHODS: We collected difficulty ratings of EHR notes and Wikipedia articles using crowdsourcing from 90 readers. We built a supervised model to assess readability based on relative orders of text difficulty using both surface text features and word embeddings. We evaluated system performance using the Kendall coefficient of concordance against human ratings. RESULTS: Our system achieved significantly higher concordance (.734) with human annotators than did a baseline using the Flesch-Kincaid Grade Level, a widely adopted readability formula (.531). The improvement was also consistent across different disease topics. This method's concordance with an individual human user's ratings was also higher than the concordance between different human annotators (.658). CONCLUSIONS: We explored methods to automatically assess the readability levels of clinical narratives. Our ranking-based system using simple textual features and easy-to-learn word embeddings outperformed a widely used readability formula. Our ranking-based method can predict relative difficulties of medical documents. It is not constrained to a predefined set of readability levels, a common design in many machine learning-based systems. Furthermore, the feature set does not rely on complex processing of the documents. One potential application of our readability ranking is personalization, allowing patients to better accommodate their own background knowledge.","journal":"The Journal of Medical Internet Research Medical Informatics","author":[{"propositions":[],"lastnames":["Zheng"],"firstnames":["JP"],"suffixes":[]},{"propositions":[],"lastnames":["Yu"],"firstnames":["H"],"suffixes":[]}],"month":"March","year":"2018","pmid":"29572199 PMCID: PMC5889493","bibtex":"@article{zheng_assessing_2018,\n\ttitle = {Assessing {Readability} of {Medical} {Documents}: {A} {Ranking} {Approach}.},\n\tdoi = {DOI: 10.2196/medinform.8611},\n\tabstract = {BACKGROUND:\nThe use of electronic health record (EHR) systems with patient engagement capabilities, including viewing, downloading, and transmitting health information, has recently grown tremendously. However, using these resources to engage patients in managing their own health remains challenging due to the complex and technical nature of the EHR narratives.\n\nOBJECTIVE:\nOur objective was to develop a machine learning-based system to assess readability levels of complex documents such as EHR notes.\n\nMETHODS:\nWe collected difficulty ratings of EHR notes and Wikipedia articles using crowdsourcing from 90 readers. We built a supervised model to assess readability based on relative orders of text difficulty using both surface text features and word embeddings. We evaluated system performance using the Kendall coefficient of concordance against human ratings.\n\nRESULTS:\nOur system achieved significantly higher concordance (.734) with human annotators than did a baseline using the Flesch-Kincaid Grade Level, a widely adopted readability formula (.531). The improvement was also consistent across different disease topics. This method's concordance with an individual human user's ratings was also higher than the concordance between different human annotators (.658).\n\nCONCLUSIONS:\nWe explored methods to automatically assess the readability levels of clinical narratives. Our ranking-based system using simple textual features and easy-to-learn word embeddings outperformed a widely used readability formula. Our ranking-based method can predict relative difficulties of medical documents. It is not constrained to a predefined set of readability levels, a common design in many machine learning-based systems. Furthermore, the feature set does not rely on complex processing of the documents. One potential application of our readability ranking is personalization, allowing patients to better accommodate their own background knowledge.},\n\tjournal = {The Journal of Medical Internet Research Medical Informatics},\n\tauthor = {Zheng, JP and Yu, H},\n\tmonth = mar,\n\tyear = {2018},\n\tpmid = {29572199 PMCID: PMC5889493},\n}\n\n","author_short":["Zheng, J.","Yu, H"],"key":"zheng_assessing_2018","id":"zheng_assessing_2018","bibbaseid":"zheng-yu-assessingreadabilityofmedicaldocumentsarankingapproach-2018","role":"author","urls":{},"metadata":{"authorlinks":{}},"html":""},"bibtype":"article","biburl":"http://fenway.cs.uml.edu/papers/pubs-all.bib","dataSources":["TqaA9miSB65nRfS5H"],"keywords":[],"search_terms":["assessing","readability","medical","documents","ranking","approach","zheng","yu"],"title":"Assessing Readability of Medical Documents: A Ranking Approach.","year":2018}