ABSTRACT
Selecting and presenting content culled from multiple heterogeneous and physically distributed sources is a challenging task. The exponential growth of the web data in modern times has brought new requirements to such integration systems. Data is not any more produced by content providers alone, but also from regular users through the highly popular Web 2.0 social and semantic web applications. The plethora of the available web content increased its demand by regular users who could not any more wait the development of advanced integration tools. They wanted to be able to build in a short time their own specialized integration applications. Aggregators came to the risk of these users. They allowed them not only to combine distributed content, but also to process it in ways that generate new services available for further consumption.
To cope with the heterogeneous data, the Linked Data initiative aims at the creation and exploitation of correspondences across data values. In this work, although we share the Linked Data community vision, we advocate that for the modern web, linking at the data value level is not enough. Aggregators should base their integration tasks on the concept of an entity, i.e., identifying whether different pieces of information correspond to the same real world entity, such as an event or a person. We describe our theory, system, and experimental results that illustrate the approach's effectiveness.
- S. Amer-Yahia, V. Markl, A. Y. Halevy, A. Doan, G. Alonso, D. Kossmann, and G. Weikum. Databases and web 2.0 panel at vldb 2007. SIGMOD Record, 2008. Google ScholarDigital Library
- P. Andritsos, A. Fuxman, and R. J. Miller. Clean answers over dirty databases: A probabilistic approach. In ICDE, 2006. Google ScholarDigital Library
- C. Bizer, T. Heath, and T. Berners-Lee. Linked Data - The story so far. IJSWIS, 2009.Google ScholarCross Ref
- N. N. Dalvi, R. Kumar, B. Pang, R. Ramakrishnan, A. Tomkins, P. Bohannon, S. Keerthi, and S. Merugu. A web of concepts. In PODS, 2009. Google ScholarDigital Library
- A. K. Elmagarmid, P. G. Ipeirotis, and V. S. Verykios. Duplicate record detection: A survey. IEEE Trans. Knowl. Data Eng., 2007. Google ScholarDigital Library
- E. Ioannou, S. Sathe, N. Bonvin, A. Jain, S. Bondalapati, G. Skobeltsyn, C. Niederée, and Z. Miklos. Entity Search with Necessity. In WebDB, 2009.Google Scholar
- G. D. Lorenzo, H. Hacid, H. young Paik, and B. Benatallah. Data integration in mashups. SIGMOD Rec., 2009. Google ScholarDigital Library
- B.-W. On, N. Koudas, D. Lee, and D. Srivastava. Group linkage. In ICDE, 2007.Google ScholarCross Ref
- M. Zhong, M. Liu, and Q. Chen. Modeling heterogeneous data in dataspace. In IEEE IRI, 2008.Google ScholarCross Ref
Index Terms
- Enabling entity-based aggregators for web 2.0 data
Recommendations
Managing Web Content Using Linked Data Principles - Combining Semantic Structure with Dynamic Content Syndication
COMPSAC '11: Proceedings of the 2011 IEEE 35th Annual Computer Software and Applications ConferenceDespite the success of the emerging Linked Data Web, offering content in a machine-processable way and--at the same time--as a traditional Web site is still not a trivial task. In this paper, we present the OntoWiki-CMS--an extension to the ...
A bootstrapping approach to entity linkage on the Semantic Web
In the Big Data era, ever-increasing RDF data have reached a scale in billions of entities and brought challenges to the problem of entity linkage on the Semantic Web. Although millions of entities, typically denoted by URIs, have been explicitly linked ...
Querying semantic web data with SPARQL
PODS '11: Proceedings of the thirtieth ACM SIGMOD-SIGACT-SIGART symposium on Principles of database systemsThe Semantic Web is the initiative of the W3C to make information on the Web readable not only by humans but also by machines. RDF is the data model for Semantic Web data, and SPARQL is the standard query language for this data model. In the last ten ...
Comments