Exploring the use of natural language systems for fact identification: Towards the automatic construction of healthcare portals
dc.contributor.author | Peck, Frederick A. | en_US |
dc.contributor.author | Bhavnani, Suresh K. | en_US |
dc.contributor.author | Blackmon, Marilyn H. | en_US |
dc.contributor.author | Radev, Dragomir R. | en_US |
dc.date.accessioned | 2006-04-19T13:40:00Z | |
dc.date.available | 2006-04-19T13:40:00Z | |
dc.date.issued | 2004 | en_US |
dc.identifier.citation | Peck, Frederick A.; Bhavnani, Suresh K.; Blackmon, Marilyn H.; Radev, Dragomir R. (2004)."Exploring the use of natural language systems for fact identification: Towards the automatic construction of healthcare portals." Proceedings of the American Society for Information Science and Technology 41(1): 327-338. <http://hdl.handle.net/2027.42/34561> | en_US |
dc.identifier.issn | 0044-7870 | en_US |
dc.identifier.issn | 1550-8390 | en_US |
dc.identifier.uri | https://hdl.handle.net/2027.42/34561 | |
dc.description.abstract | In prior work we observed that expert searchers follow well-defined search procedures in order to obtain comprehensive information on the Web. Motivated by that observation, we developed a prototype domain portal called the Strategy Hub that provides expert search procedures to benefit novice searchers. The search procedures in the prototype were entirely handcrafted by search experts, making further expansion of the Strategy Hub cost-prohibitive. However, a recent study on the distribution of healthcare information on the web suggested that search procedures can be automatically generated from pages that have been rated based on the extent to which they cover facts relevant to a topic. This paper presents the results of experiments designed to automate the process of rating the extent to which a page covers relevant facts. To automatically generate these ratings, we used two natural language systems, Latent Semantic Analysis and MEAD, to compute the similarity between sentences on the page and each fact. We then used an algorithm to convert these similarity scores to a single rating that represents the extent to which the page covered each fact. These automatic ratings are compared with manual ratings using inter-rater reliability statistics. Analysis of these statistics reveals the strengths and weaknesses of each tool, and suggests avenues for improvement. | en_US |
dc.format.extent | 1107047 bytes | |
dc.format.extent | 3118 bytes | |
dc.format.mimetype | application/pdf | |
dc.format.mimetype | text/plain | |
dc.language.iso | en_US | |
dc.publisher | Wiley Subscription Services, Inc., A Wiley Company | en_US |
dc.subject.other | Computer Science | en_US |
dc.title | Exploring the use of natural language systems for fact identification: Towards the automatic construction of healthcare portals | en_US |
dc.type | Article | en_US |
dc.rights.robots | IndexNoFollow | en_US |
dc.subject.hlbsecondlevel | Information and Library Science | en_US |
dc.subject.hlbtoplevel | Social Sciences | en_US |
dc.description.peerreviewed | Peer Reviewed | en_US |
dc.contributor.affiliationum | School of Information, University of Michigan, Ann Arbor, Ml 48109–1092 | en_US |
dc.contributor.affiliationum | School of Information, University of Michigan, Ann Arbor, Ml 48109–1092 | en_US |
dc.contributor.affiliationum | School of Information and Department of EECS, University of Michigan, Ann Arbor, Ml 48109–1092 | en_US |
dc.contributor.affiliationother | Institute of Cognitive Science, University of Colorado, Boulder, CO 80309–0344 | en_US |
dc.description.bitstreamurl | http://deepblue.lib.umich.edu/bitstream/2027.42/34561/1/1450410139_ftp.pdf | en_US |
dc.identifier.doi | http://dx.doi.org/10.1002/meet.1450410139 | en_US |
dc.identifier.source | Proceedings of the American Society for Information Science and Technology | en_US |
dc.owningcollname | Interdisciplinary and Peer-Reviewed |
Files in this item
Remediation of Harmful Language
The University of Michigan Library aims to describe library materials in a way that respects the people and communities who create, use, and are represented in our collections. Report harmful or offensive language in catalog records, finding aids, or elsewhere in our collections anonymously through our metadata feedback form. More information at Remediation of Harmful Language.
Accessibility
If you are unable to use this file in its current format, please select the Contact Us link and we can modify it to make it more accessible to you.