posted on 2009-10-01, 00:00authored byTom MitchellTom Mitchell, Justin Betteridge, Andrew Carlson, Estevan Hruschka, Richard Wang
A key question regarding the future of the semantic web is “how will we acquire structured information to populate the semantic web on a vast scale?” One approach is to enter this information manually. A second approach is to take advantage of pre-existing databases, and to develop common ontologies, publishing standards, and reward systems to make this data widely accessible. We consider here a third approach: developing software that automatically extracts structured information from unstructured text present on the web. We also describe preliminary results demonstrating that machine learning algorithms can learn to extract tens of thousands of facts to populate a diverse ontology, with imperfect but reasonably good accuracy.
History
Publisher Statement
The final publication is available at Springer via http://dx.doi.org/10.1007/978-3-642-04930-9_66