The YASGUI family of SPARQL clients.
Rietveld, L., & Hoekstra, R.
Semantic Web, 8(3): 373–383. 12 2016.
doi
link
bibtex
abstract
@article{82074b235a8a424bb7645c2c2887f457,
title = "The YASGUI family of SPARQL clients",
abstract = "The size and complexity of the Semantic Web and its technology stack makes it difficult to query. Access to Linked Data could be greatly facilitated if it were supported by a tool with a strong focus on usability. In this paper we present the YASGUI family of SPARQL clients, a continuation of the YASGUI tool introduced more than two years ago. The YASGUI family of SPARQL clients enables publishers to improve ease of access for their SPARQL endpoints, and gives consumers of Linked Data a robust, feature-rich and user friendly SPARQL editor. We show that the YASGUI family had significant impact on the landscape of Linked Data management: YASGUI components are integrated in state-of-the-art triple-stores and Linked Data applications, and used as front-end by a large number of Linked Data publishers. Additionally, we show that the YASGUI web service - which provides access to any SPARQL endpoint - has a large and growing user base amongst Linked Data consumers.",
keywords = "data publishing, Linked Data, query formulation, SPARQL",
author = "Laurens Rietveld and Rinke Hoekstra",
year = "2016",
month = "12",
doi = "10.3233/SW-150197",
volume = "8",
pages = "373--383",
journal = "Semantic Web",
issn = "1570-0844",
publisher = "IOS Press",
number = "3",
}
The size and complexity of the Semantic Web and its technology stack makes it difficult to query. Access to Linked Data could be greatly facilitated if it were supported by a tool with a strong focus on usability. In this paper we present the YASGUI family of SPARQL clients, a continuation of the YASGUI tool introduced more than two years ago. The YASGUI family of SPARQL clients enables publishers to improve ease of access for their SPARQL endpoints, and gives consumers of Linked Data a robust, feature-rich and user friendly SPARQL editor. We show that the YASGUI family had significant impact on the landscape of Linked Data management: YASGUI components are integrated in state-of-the-art triple-stores and Linked Data applications, and used as front-end by a large number of Linked Data publishers. Additionally, we show that the YASGUI web service - which provides access to any SPARQL endpoint - has a large and growing user base amongst Linked Data consumers.
.
Wang, H., Huang, Z., Zhong, N., Huang, J., Han, Y., & Zhang, F.
A Monitoring System for the Safety of Building Structure Based on W2T Methodology, pages 323–335. Zhong, N., Ma, J., Liu, J., Huang, R., & Tao, X., editor(s). Springer, 11 2016.
doi
link
bibtex
abstract
@inbook{6d083df218bb468baf4d48b4dc55be0b,
title = "A Monitoring System for the Safety of Building Structure Based on W2T Methodology",
abstract = "With the development of the Internet of things, monitoring systems for the safety of building structure (SBS) provide people with the important data about the main supporting points in the buildings. More and more data give the engineers an overload work problem, which can be solved by a systematic method making these monitoring systems more reliable, efficient and intelligent. Under the framework of the Wisdom Web of Things (W2T), we design a monitoring system for the SBS, by using the semantic technology. This system establishes a data cycle among the physical world (buildings), the social world (humans) and the cyber world (computers), and provides various services in the monitoring process to alleviate the engineers’ workload. In this system, the sensors which are connected via cable or wireless way, are used to monitor the different parameters of building structure. The semantic data can be obtained and represented by RDF to describe the meanings of sensor data, and can provide the application background for users. LarKC, a platform for scalable semantic data processing, is used for semantic querying about the data. Based on this uniform representation of data and semantic processing, intelligent services can be provided by the effective data analysis. This provides the possibility to integrate all of the monitoring systems for the safety of building structure in urban computing.",
author = "Haiyuan Wang and Zhisheng Huang and Ning Zhong and Jiajin Huang and Yuzhong Han and Feng Zhang",
year = "2016",
month = "11",
doi = "10.1007/978-3-319-44198-6",
isbn = "978-3-319-44196-2",
pages = "323--335",
editor = "Ning Zhong and Jianhua Ma and Jiming Liu and Runhe Huang and Xiaohui Tao",
booktitle = "Wisdom Web of Things",
publisher = "Springer",
}
With the development of the Internet of things, monitoring systems for the safety of building structure (SBS) provide people with the important data about the main supporting points in the buildings. More and more data give the engineers an overload work problem, which can be solved by a systematic method making these monitoring systems more reliable, efficient and intelligent. Under the framework of the Wisdom Web of Things (W2T), we design a monitoring system for the SBS, by using the semantic technology. This system establishes a data cycle among the physical world (buildings), the social world (humans) and the cyber world (computers), and provides various services in the monitoring process to alleviate the engineers’ workload. In this system, the sensors which are connected via cable or wireless way, are used to monitor the different parameters of building structure. The semantic data can be obtained and represented by RDF to describe the meanings of sensor data, and can provide the application background for users. LarKC, a platform for scalable semantic data processing, is used for semantic querying about the data. Based on this uniform representation of data and semantic processing, intelligent services can be provided by the effective data analysis. This provides the possibility to integrate all of the monitoring systems for the safety of building structure in urban computing.
.
Wang, H., Huang, J., Zhou, E., Huang, Z., & Zhong, N.
Suitable Route Recommendation Inspired by Cognition, pages 303–322. Zhong, N., Ma, J., Liu, J., Huang, R., & Tao, X., editor(s). Springer, 11 2016.
doi
link
bibtex
abstract
@inbook{66b040c64f254e379e952c8ef535213a,
title = "Suitable Route Recommendation Inspired by Cognition",
abstract = "With the increasing popularity of mobile phones, large amounts of real and reliable mobile phone data are being generated every day. These mobile phone data represent the practical travel routes of users and imply the intelligence of them in selecting a suitable route. Usually, an experienced user knows which route is congested in a specified period of time but unblocked in another period of time. Moreover, a route used frequently and recently by a user is usually the suitable one to satisfy the user’s needs. ACT-R (Adaptive Control of Thought-Rational) is a computational cognitive architecture, which provides a good framework to understand the principles and mechanisms of information organization, retrieval and selection in human memory. In this chapter, we employ ACT-R to model the process of selecting a suitable route of users. We propose a cognition-inspired route recommendation method to mine the intelligence of users in selecting a suitable route, evaluate the suitability of the routes, and recommend an ordered list of routes for subscribers. Experiments show that it is effective and feasible to recommend the suitable routes inspired by cognition.",
author = "Hui Wang and Jiajin Huang and Erzhong Zhou and Zhisheng Huang and Ning Zhong",
year = "2016",
month = "11",
doi = "10.1007/978-3-319-44198-6_13",
isbn = "978-3-319-44196-2",
pages = "303--322",
editor = "Ning Zhong and Jianhua Ma and Jiming Liu and Runhe Huang and Xiaohui Tao",
booktitle = "Wisdom Web of Things",
publisher = "Springer",
}
With the increasing popularity of mobile phones, large amounts of real and reliable mobile phone data are being generated every day. These mobile phone data represent the practical travel routes of users and imply the intelligence of them in selecting a suitable route. Usually, an experienced user knows which route is congested in a specified period of time but unblocked in another period of time. Moreover, a route used frequently and recently by a user is usually the suitable one to satisfy the user’s needs. ACT-R (Adaptive Control of Thought-Rational) is a computational cognitive architecture, which provides a good framework to understand the principles and mechanisms of information organization, retrieval and selection in human memory. In this chapter, we employ ACT-R to model the process of selecting a suitable route of users. We propose a cognition-inspired route recommendation method to mine the intelligence of users in selecting a suitable route, evaluate the suitability of the routes, and recommend an ordered list of routes for subscribers. Experiments show that it is effective and feasible to recommend the suitable routes inspired by cognition.
.
Chen, J., Ma, J., Zhong, N., Yao, Y., Liu, J., Huang, R., Li, W., Huang, Z., & Gao, Y.
WaaS—Wisdom as a Service, pages 27–46. Zhong, N., Ma, J., Liu, J., Huang, R., & Tao, X., editor(s). 11 2016.
doi
link
bibtex
abstract
@inbook{cdce099b6d134e23a733033930db4362,
title = "WaaS—Wisdom as a Service",
abstract = "An emerging hyper-world encompasses all human activities in a social-cyber-physical space. Its power derives from the Wisdom Web of Things (W2T) cycle, namely, “from things to data, information, knowledge, wisdom, services, humans, and then back to things.” The W2T cycle leads to a harmonious symbiosis among humans, computers and things, which can be constructed by large-scale converging of intelligent information technology applications with an open and interoperable architecture. The recent advances in cloud computing, the Internet/Web of Things, big data and other research fields have provided just such an open system architecture with resource sharing/services. The next step is therefore to develop an open and interoperable content architecture with intelligence sharing/services for the organization and transformation in the “data, information, knowledge and wisdom (DIKW)” hierarchy. This chapter introduces Wisdom as a Service (WaaS) as a content architecture based on the “paying only for what you use” IT business trend. The WaaS infrastructure, WaaS economics, and the main challenges in WaaS research and applications are discussed. A case study is described to demonstrate the usefulness and significance of WaaS. Relying on the clouds (cloud computing), things (Internet of Things) and big data, WaaS provides a practical approach to realize the W2T cycle in the hyper-world for the coming age of ubiquitous intelligent IT applications.",
author = "Jianhui Chen and Jianhua Ma and Ning Zhong and Yiyu Yao and Jiming Liu and Runhe Huang and Wenbin Li and Zhisheng Huang and Yang Gao",
year = "2016",
month = "11",
doi = "10.1007/978-3-319-44198-6_2",
isbn = "978-3-319-44196-2",
pages = "27--46",
editor = "Ning Zhong and Jianhua Ma and Jiming Liu and Runhe Huang and Xiaohui Tao",
booktitle = "Wisdom Web of Things",
}
An emerging hyper-world encompasses all human activities in a social-cyber-physical space. Its power derives from the Wisdom Web of Things (W2T) cycle, namely, “from things to data, information, knowledge, wisdom, services, humans, and then back to things.” The W2T cycle leads to a harmonious symbiosis among humans, computers and things, which can be constructed by large-scale converging of intelligent information technology applications with an open and interoperable architecture. The recent advances in cloud computing, the Internet/Web of Things, big data and other research fields have provided just such an open system architecture with resource sharing/services. The next step is therefore to develop an open and interoperable content architecture with intelligence sharing/services for the organization and transformation in the “data, information, knowledge and wisdom (DIKW)” hierarchy. This chapter introduces Wisdom as a Service (WaaS) as a content architecture based on the “paying only for what you use” IT business trend. The WaaS infrastructure, WaaS economics, and the main challenges in WaaS research and applications are discussed. A case study is described to demonstrate the usefulness and significance of WaaS. Relying on the clouds (cloud computing), things (Internet of Things) and big data, WaaS provides a practical approach to realize the W2T cycle in the hyper-world for the coming age of ubiquitous intelligent IT applications.
Linked Data for Digital History: Lessons Learned from Three Case Studies.
de Boer , V., Merono Penuela, A., & Ockeloen, C.
Anejos de la Revista de Historiografía, (4): 139–162. 10 2016.
link
bibtex
@article{22ff35d5e5d7447b852af14f2bf7aba5,
title = "Linked Data for Digital History: Lessons Learned from Three Case Studies",
author = "{de Boer}, V. and {Merono Penuela}, A. and C.J. Ockeloen",
year = "2016",
month = "10",
pages = "139--162",
journal = "Anejos de la Revista de Historiografía",
number = "4",
}
Special Issue: Modern Hardware.
Boncz, P., Lehner, W., & Neumann, T.
VLDB Journal, 25(5): 623–624. 10 2016.
doi
link
bibtex
@article{a4774a07abff41b98843780ec580fafc,
title = "Special Issue: Modern Hardware",
author = "Peter Boncz and Wolfgang Lehner and Thomas Neumann",
year = "2016",
month = "10",
doi = "10.1007/s00778-016-0440-7",
volume = "25",
pages = "623--624",
journal = "VLDB Journal",
issn = "1066-8888",
publisher = "Springer Verlag",
number = "5",
}
Predictive modeling of colorectal cancer using a dedicated pre-processing pipeline on routine electronic medical records.
Kop, R., Hoogendoorn, M., ten Teije , A., Büchner, F., Slottje, P., Moons, L., & Numans, M.
Computers in Biology and Medicine, 76: 30–38. 9 2016.
doi
link
bibtex
abstract
@article{8fcbb631b38f4a52999735c46e47434a,
title = "Predictive modeling of colorectal cancer using a dedicated pre-processing pipeline on routine electronic medical records",
abstract = "Over the past years, research utilizing routine care data extracted from Electronic Medical Records (EMRs) has increased tremendously. Yet there are no straightforward, standardized strategies for pre-processing these data. We propose a dedicated medical pre-processing pipeline aimed at taking on many problems and opportunities contained within EMR data, such as their temporal, inaccurate and incomplete nature. The pipeline is demonstrated on a dataset of routinely recorded data in general practice EMRs of over 260,000 patients, in which the occurrence of colorectal cancer (CRC) is predicted using various machine learning techniques (i.e., CART, LR, RF) and subsets of the data. CRC is a common type of cancer, of which early detection has proven to be important yet challenging. The results are threefold. First, the predictive models generated using our pipeline reconfirmed known predictors and identified new, medically plausible, predictors derived from the cardiovascular and metabolic disease domain, validating the pipeline's effectiveness. Second, the difference between the best model generated by the data-driven subset (AUC 0.891) and the best model generated by the current state of the art hypothesis-driven subset (AUC 0.864) is statistically significant at the 95% confidence interval level. Third, the pipeline itself is highly generic and independent of the specific disease targeted and the EMR used. In conclusion, the application of established machine learning techniques in combination with the proposed pipeline on EMRs has great potential to enhance disease prediction, and hence early detection and intervention in medical practice.",
keywords = "Colorectal cancer, Data mining, Data processing, Electronic medical records, Machine learning",
author = "Reinier Kop and Mark Hoogendoorn and {ten Teije}, Annette and Büchner, {Frederike L.} and Pauline Slottje and Moons, {Leon M G} and Numans, {Mattijs E.}",
year = "2016",
month = "9",
doi = "10.1016/j.compbiomed.2016.06.019",
volume = "76",
pages = "30--38",
journal = "Computers in Biology and Medicine",
issn = "0010-4825",
publisher = "Elsevier Limited",
}
Over the past years, research utilizing routine care data extracted from Electronic Medical Records (EMRs) has increased tremendously. Yet there are no straightforward, standardized strategies for pre-processing these data. We propose a dedicated medical pre-processing pipeline aimed at taking on many problems and opportunities contained within EMR data, such as their temporal, inaccurate and incomplete nature. The pipeline is demonstrated on a dataset of routinely recorded data in general practice EMRs of over 260,000 patients, in which the occurrence of colorectal cancer (CRC) is predicted using various machine learning techniques (i.e., CART, LR, RF) and subsets of the data. CRC is a common type of cancer, of which early detection has proven to be important yet challenging. The results are threefold. First, the predictive models generated using our pipeline reconfirmed known predictors and identified new, medically plausible, predictors derived from the cardiovascular and metabolic disease domain, validating the pipeline's effectiveness. Second, the difference between the best model generated by the data-driven subset (AUC 0.891) and the best model generated by the current state of the art hypothesis-driven subset (AUC 0.864) is statistically significant at the 95% confidence interval level. Third, the pipeline itself is highly generic and independent of the specific disease targeted and the EMR used. In conclusion, the application of established machine learning techniques in combination with the proposed pipeline on EMRs has great potential to enhance disease prediction, and hence early detection and intervention in medical practice.
.
Costea, A., Ionescu, A., Raducanu, B., Świtakowski, M., Bârca, C., Sompolski, J., Łuszczak, A., Szafrański, M., De Nijs, G., & Boncz, P.
Volume 26-June-2016 . VectorH: Taking SQL-on-Hadoop to the next level, pages 1105–1117. Association for Computing Machinery (ACM), 6 2016.
doi
link
bibtex
abstract
@inbook{81f07f72c43d466a8c80eb41d581cc13,
title = "VectorH: Taking SQL-on-Hadoop to the next level",
abstract = "Actian Vector in Hadoop (VectorH for short) is a new SQL-on-Hadoop system built on top of the fast Vectorwise analytical database system. VectorH achieves fault tolerance and storage scalability by relying on HDFS, and extends the state-of-the-art in SQL-on-Hadoop systems by instrumenting the HDFS replication policy to optimize read locality. VectorH integrates with YARN for workload management, achieving a high degree of elasticity. Even though HDFS is an append-only file-system, and VectorH supports (update-averse) ordered tables, trickle updates are possible thanks to Positional Delta Trees (PDTs), a diffferential update structure that can be queried efficiently. We describe the changes made to single-server Vectorwise to turn it into a Hadoop-based MPP system, encompassing workload management, parallel query optimization and execution, HDFS storage, transaction processing and Spark integration. We evaluate VectorH against HAWQ, Impala, SparkSQL and Hive, showing orders of magnitude better performance.",
author = "Andrei Costea and Adrian Ionescu and Bogdan Raducanu and Michał Świtakowski and Cristian Bârca and Juliusz Sompolski and Alicja Łuszczak and Michał Szafrański and {De Nijs}, Giel and Peter Boncz",
year = "2016",
month = "6",
doi = "10.1145/2882903.2903742",
volume = "26-June-2016",
pages = "1105--1117",
booktitle = "SIGMOD 2016 - Proceedings of the 2016 International Conference on Management of Data",
publisher = "Association for Computing Machinery (ACM)",
}
Actian Vector in Hadoop (VectorH for short) is a new SQL-on-Hadoop system built on top of the fast Vectorwise analytical database system. VectorH achieves fault tolerance and storage scalability by relying on HDFS, and extends the state-of-the-art in SQL-on-Hadoop systems by instrumenting the HDFS replication policy to optimize read locality. VectorH integrates with YARN for workload management, achieving a high degree of elasticity. Even though HDFS is an append-only file-system, and VectorH supports (update-averse) ordered tables, trickle updates are possible thanks to Positional Delta Trees (PDTs), a diffferential update structure that can be queried efficiently. We describe the changes made to single-server Vectorwise to turn it into a Hadoop-based MPP system, encompassing workload management, parallel query optimization and execution, HDFS storage, transaction processing and Spark integration. We evaluate VectorH against HAWQ, Impala, SparkSQL and Hive, showing orders of magnitude better performance.
.
Rutgers, P., Martella, C., Voulgaris, S., & Boncz, P.
Volume 24-June-2016 . Powerful and efficient bulk shortest-path queries: Cypher language extension & Giraph implementation. 6 2016.
doi
link
bibtex
@inbook{4c8c9c9bb95a4376a43d9eaa67f7db00,
title = "Powerful and efficient bulk shortest-path queries: Cypher language extension & Giraph implementation",
author = "Peter Rutgers and Claudio Martella and Spyros Voulgaris and Peter Boncz",
year = "2016",
month = "6",
doi = "10.1145/2960414.2960420",
volume = "24-June-2016",
booktitle = "ACM International Conference Proceeding Series",
}
ClioPatria: A SWI-prolog infrastructure for the semantic web.
Wielemaker, J., Beek, W., Hildebrand, M., & Van Ossenbruggen, J.
Semantic Web, 7(5): 529–541. 6 2016.
doi
link
bibtex
abstract
@article{d35101d170504aa89903f06633f2236d,
title = "ClioPatria: A SWI-prolog infrastructure for the semantic web",
abstract = "ClioPatria is a comprehensive semantic web development framework based on SWI-Prolog. SWI-Prolog provides an efficient C-based main-memory RDF store that is designed to cooperate naturally and efficiently with Prolog, realizing a flexible RDF-based environment for rule based programming. ClioPatria extends this core with a SPARQL and LOD server, an extensible web frontend to manage the server, browse the data, query the data using SPARQL and Prolog and a Git-based plugin manager. The ability to query RDF using Prolog provides query composition and smooth integration with application logic. ClioPatria is primarily positioned as a prototyping platform for exploring novel ways of reasoning with RDF data. It has been used in several research projects in order to perform tasks such as data integration and enrichment and semantic search.",
keywords = "Logic programming, Semantic Web framework, Triple store",
author = "Jan Wielemaker and Wouter Beek and Michiel Hildebrand and {Van Ossenbruggen}, Jacco",
year = "2016",
month = "6",
doi = "10.3233/SW-150191",
volume = "7",
pages = "529--541",
journal = "Semantic Web",
issn = "1570-0844",
publisher = "IOS Press",
number = "5",
}
ClioPatria is a comprehensive semantic web development framework based on SWI-Prolog. SWI-Prolog provides an efficient C-based main-memory RDF store that is designed to cooperate naturally and efficiently with Prolog, realizing a flexible RDF-based environment for rule based programming. ClioPatria extends this core with a SPARQL and LOD server, an extensible web frontend to manage the server, browse the data, query the data using SPARQL and Prolog and a Git-based plugin manager. The ability to query RDF using Prolog provides query composition and smooth integration with application logic. ClioPatria is primarily positioned as a prototyping platform for exploring novel ways of reasoning with RDF data. It has been used in several research projects in order to perform tasks such as data integration and enrichment and semantic search.
Bitwise dimensional co-clustering for analytical workloads.
Baumann, S., Boncz, P., & Sattler, K.
VLDB Journal, 25(3): 291–316. 6 2016.
doi
link
bibtex
abstract
@article{e1ddb0419e3f41a181b986d27ca0553e,
title = "Bitwise dimensional co-clustering for analytical workloads",
abstract = "Analytical workloads in data warehouses often include heavy joins where queries involve multiple fact tables in addition to the typical star-patterns, dimensional grouping and selections. In this paper we propose a new processing and storage framework called bitwise dimensional co-clustering (BDCC) that avoids replication and thus keeps updates fast, yet is able to accelerate all these foreign key joins, efficiently support grouping and pushes down most dimensional selections. The core idea of BDCC is to cluster each table on a mix of dimensions, each possibly derived from attributes imported over an incoming foreign key and this way creating foreign key connected tables with partially shared clusterings. These are later used to accelerate any join between two tables that have some dimension in common and additionally permit to push down and propagate selections (reduce I/O) and accelerate aggregation and ordering operations. Besides the general framework, we describe an algorithm to derive such a physical co-clustering database automatically and describe query processing and query optimization techniques that can easily be fitted into existing relational engines. We present an experimental evaluation on the TPC-H benchmark in the Vectorwise system, showing that co-clustering can significantly enhance its already high performance and at the same time significantly reduce the memory consumption of the system.",
keywords = "Clustering, Data warehouse, Database design, Indexing, OLAP, Query processing, Sandwich operators, Storage",
author = "Stephan Baumann and Peter Boncz and Sattler, {Kai Uwe}",
year = "2016",
month = "6",
doi = "10.1007/s00778-015-0417-y",
volume = "25",
pages = "291--316",
journal = "VLDB Journal",
issn = "1066-8888",
publisher = "Springer Verlag",
number = "3",
}
Analytical workloads in data warehouses often include heavy joins where queries involve multiple fact tables in addition to the typical star-patterns, dimensional grouping and selections. In this paper we propose a new processing and storage framework called bitwise dimensional co-clustering (BDCC) that avoids replication and thus keeps updates fast, yet is able to accelerate all these foreign key joins, efficiently support grouping and pushes down most dimensional selections. The core idea of BDCC is to cluster each table on a mix of dimensions, each possibly derived from attributes imported over an incoming foreign key and this way creating foreign key connected tables with partially shared clusterings. These are later used to accelerate any join between two tables that have some dimension in common and additionally permit to push down and propagate selections (reduce I/O) and accelerate aggregation and ordering operations. Besides the general framework, we describe an algorithm to derive such a physical co-clustering database automatically and describe query processing and query optimization techniques that can easily be fitted into existing relational engines. We present an experimental evaluation on the TPC-H benchmark in the Vectorwise system, showing that co-clustering can significantly enhance its already high performance and at the same time significantly reduce the memory consumption of the system.
How organisation of architecture documentation affects architectural knowledge retrieval.
de Graaf , K., Liang, P., Tang, A., & Vliet, J.
Science of Computer Programming, 121: 75–99. 6 2016.
doi
link
bibtex
abstract
@article{283d4b8ef23a46bb9ecc246cfc723196,
title = "How organisation of architecture documentation affects architectural knowledge retrieval",
abstract = "A common approach to software architecture documentation in industry projects is the use of file-based documents. This approach offers a single-dimensional arrangement of the architectural knowledge. Knowledge retrieval from file-based architecture documentation is efficient if the organisation of knowledge supports the needs of the readers; otherwise it can be difficult. In this paper, we compare the organisation and retrieval of architectural knowledge in a file-based documentation approach and an ontology-based documentation approach. The ontology-based approach offers a multi-dimensional organisation of architectural knowledge by means of a software ontology and semantic wiki, whereas file-based documentation typically uses hierarchical organisation by directory structure and table of content. We conducted case studies in two companies to study the efficiency and effectiveness of retrieving architectural knowledge from the different organisations of knowledge. We found that the use of better knowledge organisation correlates with the efficiency and effectiveness of AK retrieval. Professionals who used the knowledge organisation found this beneficial.",
keywords = "Software architecture documentation, software architectural knowledge, architectural knowledge retrieval, software ontologies, semantic wiki, ontology-based documentation",
author = "{de Graaf}, K.A. and P. Liang and A. Tang and J.C. Vliet",
year = "2016",
month = "6",
doi = "10.1016/j.scico.2015.10.014",
volume = "121",
pages = "75--99",
journal = "Science of Computer Programming",
issn = "0167-6423",
publisher = "Elsevier",
}
A common approach to software architecture documentation in industry projects is the use of file-based documents. This approach offers a single-dimensional arrangement of the architectural knowledge. Knowledge retrieval from file-based architecture documentation is efficient if the organisation of knowledge supports the needs of the readers; otherwise it can be difficult. In this paper, we compare the organisation and retrieval of architectural knowledge in a file-based documentation approach and an ontology-based documentation approach. The ontology-based approach offers a multi-dimensional organisation of architectural knowledge by means of a software ontology and semantic wiki, whereas file-based documentation typically uses hierarchical organisation by directory structure and table of content. We conducted case studies in two companies to study the efficiency and effectiveness of retrieving architectural knowledge from the different organisations of knowledge. We found that the use of better knowledge organisation correlates with the efficiency and effectiveness of AK retrieval. Professionals who used the knowledge organisation found this beneficial.
LOD Laundromat: Why the Semantic Web needs centralization (even if we don't like it).
Beek, W., Rietveld, L., Schlobach, S., & van Harmelen , F.
IEEE Internet Computing, 20(2): 78–81. 3 2016.
doi
link
bibtex
abstract
@article{9254330a91e749cdb89b65db174225b4,
title = "LOD Laundromat: Why the Semantic Web needs centralization (even if we don't like it)",
abstract = "LOD Laundromat poses a centralized solution for today's Semantic Web problems. This approach adheres more closely to the original vision of a Web of Data, providing uniform access to a large and ever-increasing subcollection of the LOD Cloud.",
keywords = "Internet/Web technologies, Linked Data, Linked Open Data, LOD, Semantic Web",
author = "Wouter Beek and Laurens Rietveld and Stefan Schlobach and {van Harmelen}, Frank",
year = "2016",
month = "3",
doi = "10.1109/MIC.2016.43",
volume = "20",
pages = "78--81",
journal = "IEEE Internet Computing",
issn = "1089-7801",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
number = "2",
}
LOD Laundromat poses a centralized solution for today's Semantic Web problems. This approach adheres more closely to the original vision of a Web of Data, providing uniform access to a large and ever-increasing subcollection of the LOD Cloud.
.
Beek, W., Schlobach, K., & van Harmelen , F.
Volume 9678 of Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). A Contextualised Semantics for owl: sameAs, pages 405–419. Springer/Verlag, 2016.
doi
link
bibtex
abstract
1 download
@inbook{e7e1e9893fe14b078edbf1779f4b1c68,
title = "A Contextualised Semantics for owl: sameAs",
abstract = "Identity relations are at the foundation of the Semantic Web and the Linked Data Cloud. In many instances the classical interpretation of identity is too strong for practical purposes. This is particularly the case when two entities are considered the same in some but not all contexts. Unfortunately, modeling the specific contexts in which an identity relation holds is cumbersome and, due to arbitrary reuse and the Open World Assumption, it is impossible to anticipate all contexts in which an entity will be used. We propose an alternative semantics for owl:sameAs that partitions the original relation into a hierarchy of subrelations. The subrelation to which an identity statement belongs depends on the dataset in which the statement occurs. Adding future assertions may change the subrelation to which an identity statement belongs, resulting in a context-dependent and non-monotonic semantics. We show that this more fine-grained semantics is better able to characterize the actual use of owl:sameAs as observed in Linked Open Datasets.",
author = "W.G.J. Beek and K.S. Schlobach and {van Harmelen}, F.A.H.",
year = "2016",
doi = "10.1007/978-3-319-34129-3_25",
isbn = "9783319341286",
volume = "9678",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
publisher = "Springer/Verlag",
pages = "405--419",
booktitle = "The Semantic Web. Latest Advances and New Domains - 13th International Conference, ESWC 2016, Heraklion, Crete, Greece, May 29 - June 2, 2016, Proceedings",
}
Identity relations are at the foundation of the Semantic Web and the Linked Data Cloud. In many instances the classical interpretation of identity is too strong for practical purposes. This is particularly the case when two entities are considered the same in some but not all contexts. Unfortunately, modeling the specific contexts in which an identity relation holds is cumbersome and, due to arbitrary reuse and the Open World Assumption, it is impossible to anticipate all contexts in which an entity will be used. We propose an alternative semantics for owl:sameAs that partitions the original relation into a hierarchy of subrelations. The subrelation to which an identity statement belongs depends on the dataset in which the statement occurs. Adding future assertions may change the subrelation to which an identity statement belongs, resulting in a context-dependent and non-monotonic semantics. We show that this more fine-grained semantics is better able to characterize the actual use of owl:sameAs as observed in Linked Open Datasets.
.
Khalili, A., & van den Besselaar , P.
Adaptive delineation of functional urban areas using Linked Open Data., pages 72. SNI, 2016.
link
bibtex
@inbook{3545ab263cc3471aace89f0a628dd12d,
title = "Adaptive delineation of functional urban areas using Linked Open Data.",
author = "A. Khalili and {van den Besselaar}, P.A.A.",
year = "2016",
pages = "72",
booktitle = "RISIS Deliverable 9.2",
publisher = "SNI",
}
.
Khalili, A., Loizou, A., & van Harmelen , F.
Volume 9678 of Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). Adaptive Linked Data-Driven Web Components: Building Flexible and Reusable Semantic Web Interfaces - Building Flexible and Reusable Semantic Web Interfaces, pages 677–692. Springer, 2016.
doi
link
bibtex
abstract
@inbook{cec94fa27b524331a0c5f29e4d4c28c9,
title = "Adaptive Linked Data-Driven Web Components: Building Flexible and Reusable Semantic Web Interfaces - Building Flexible and Reusable Semantic Web Interfaces",
abstract = "Due to the increasing amount of Linked Data openly published on the Web, user-facing Linked Data Applications (LDAs) are gaining momentum. One of the major entrance barriers for Web developers to contribute to this wave of LDAs is the required knowledge of Semantic Web (SW) technologies such as the RDF data model and SPARQL query language. This paper presents an adaptive component-based approach together with its open source implementation for creating flexible and reusable SW interfaces driven by Linked Data. Linked Data-driven (LD-R) Web components abstract the complexity of the underlying SW technologies in order to allow reuse of existing Web components in LDAs, enabling Web developers who are not experts in SW to develop interfaces that view, edit and browse Linked Data. In addition to the modularity provided by the LD-R components, the proposed RDF-based configuration method allows application assemblers to reshape their user interface for different use cases, by either reusing existing shared configurations or by creating their proprietary configurations.",
author = "Ali Khalili and A. Loizou and {van Harmelen}, F.A.H.",
year = "2016",
doi = "10.1007/978-3-319-34129-3_41",
isbn = "978-3-319-34128-6",
volume = "9678",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
publisher = "Springer",
pages = "677--692",
booktitle = "The Semantic Web. Latest Advances and New Domains - 13th International Conference, ESWC 2016, Heraklion, Crete, Greece, May 29 - June 2, 2016, Proceedings",
}
Due to the increasing amount of Linked Data openly published on the Web, user-facing Linked Data Applications (LDAs) are gaining momentum. One of the major entrance barriers for Web developers to contribute to this wave of LDAs is the required knowledge of Semantic Web (SW) technologies such as the RDF data model and SPARQL query language. This paper presents an adaptive component-based approach together with its open source implementation for creating flexible and reusable SW interfaces driven by Linked Data. Linked Data-driven (LD-R) Web components abstract the complexity of the underlying SW technologies in order to allow reuse of existing Web components in LDAs, enabling Web developers who are not experts in SW to develop interfaces that view, edit and browse Linked Data. In addition to the modularity provided by the LD-R components, the proposed RDF-based configuration method allows application assemblers to reshape their user interface for different use cases, by either reusing existing shared configurations or by creating their proprietary configurations.
A Deep Neural Network for Link Prediction on Knowledge Graphs.
Wilcke, W., de Boer , V., van Harmelen , F., & de Kleijn , M.
2016.
link
bibtex
@misc{ca63a27c612148d2bf29a6b1283c8e58,
title = "A Deep Neural Network for Link Prediction on Knowledge Graphs",
author = "W.X. Wilcke and {de Boer}, V. and {van Harmelen}, F.A.H. and {de Kleijn}, M.T.M.",
year = "2016",
}
An ecosystem for Linked Humanities Data.
Hoekstra, R., Meroño-Peñuela, A., Dentler, K., Rijpma, A., Zijdeman, R., & Zandhuis, I.
CEUR workshop proceedings, 1608: 85–96. 2016.
doi
link
bibtex
abstract
@article{e3f25d50b5a44b00a2906e5dd944c2a9,
title = "An ecosystem for Linked Humanities Data",
abstract = "The main promise of the digital humanities is the ability to perform scholar studies at a much broader scale, and in a much more reusable fashion. The key enabler for such studies is the availability of sufficiently well described data. For the field of socio-economic history, data usually comes in a tabular form. Existing efforts to curate and publish datasets take a top-down approach and are focused on large collections. This paper presents QBer and the underlying structured datahub, which address the long tail of research data by catering for the needs of individual scholars. QBer allows researchers to publish their (small) datasets, link them to existing vocabularies and other datasets, and thereby contribute to a growing collection of interlinked datasets. We present QBer, and evaluate our first results by showing how our system facilitates two use cases in socio-economic history.",
keywords = "digital humanities, structured data, linked data, QBer",
author = "Rinke Hoekstra and Albert Meroño-Peñuela and Kathrin Dentler and Auke Rijpma and Richard Zijdeman and Ivo Zandhuis",
year = "2016",
doi = "10.1007/978-3-319-47602-5_54",
volume = "1608",
pages = "85--96",
journal = "CEUR workshop proceedings",
issn = "1613-0073",
publisher = "CEUR Workshop Proceedings",
}
The main promise of the digital humanities is the ability to perform scholar studies at a much broader scale, and in a much more reusable fashion. The key enabler for such studies is the availability of sufficiently well described data. For the field of socio-economic history, data usually comes in a tabular form. Existing efforts to curate and publish datasets take a top-down approach and are focused on large collections. This paper presents QBer and the underlying structured datahub, which address the long tail of research data by catering for the needs of individual scholars. QBer allows researchers to publish their (small) datasets, link them to existing vocabularies and other datasets, and thereby contribute to a growing collection of interlinked datasets. We present QBer, and evaluate our first results by showing how our system facilitates two use cases in socio-economic history.
.
de Rooij , S., Beek, W., Bloem, P., van Harmelen , F., & Schlobach, S.
Volume 9981 LNCS of Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). Are names meaningful? Quantifying social meaning on the semantic web, pages 184–199. Springer/Verlag, 2016.
doi
link
bibtex
abstract
@inbook{bd128ebb57df4bfdb31cc455c81cdbc6,
title = "Are names meaningful? Quantifying social meaning on the semantic web",
abstract = "According to its model-theoretic semantics, Semantic Web IRIs are individual constants or predicate letters whose names are chosen arbitrarily and carry no formal meaning. At the same time it is a well-known aspect of Semantic Web pragmatics that IRIs are often constructed mnemonically, in order to be meaningful to a human interpreter. The latter has traditionally been termed ‘social meaning’, a concept that has been discussed but not yet quantitatively studied by the Semantic Web community. In this paper we use measures of mutual information content and methods from statistical model learning to quantify the meaning that is (at least) encoded in Semantic Web names. We implement the approach and evaluate it over hundreds of thousands of datasets in order to illustrate its efficacy. Our experiments confirm that many Semantic Web names are indeed meaningful and, more interestingly, we provide a quantitative lower bound on how much meaning is encoded in names on a per-dataset basis. To our knowledge, this is the first paper about the interaction between social and formal meaning, as well as the first paper that uses statistical model learning as a method to quantify meaning in the Semantic Web context. These insights are useful for the design of a new generation of Semantic Web tools that take such social meaning into account.",
author = "{de Rooij}, Steven and Wouter Beek and Peter Bloem and {van Harmelen}, Frank and Stefan Schlobach",
year = "2016",
doi = "10.1007/978-3-319-46523-4_12",
isbn = "978-3-319-46522-7",
volume = "9981 LNCS",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
publisher = "Springer/Verlag",
pages = "184--199",
booktitle = "The Semantic Web - 15th International Semantic Web Conference, ISWC 2016, Proceedings",
}
According to its model-theoretic semantics, Semantic Web IRIs are individual constants or predicate letters whose names are chosen arbitrarily and carry no formal meaning. At the same time it is a well-known aspect of Semantic Web pragmatics that IRIs are often constructed mnemonically, in order to be meaningful to a human interpreter. The latter has traditionally been termed ‘social meaning’, a concept that has been discussed but not yet quantitatively studied by the Semantic Web community. In this paper we use measures of mutual information content and methods from statistical model learning to quantify the meaning that is (at least) encoded in Semantic Web names. We implement the approach and evaluate it over hundreds of thousands of datasets in order to illustrate its efficacy. Our experiments confirm that many Semantic Web names are indeed meaningful and, more interestingly, we provide a quantitative lower bound on how much meaning is encoded in names on a per-dataset basis. To our knowledge, this is the first paper about the interaction between social and formal meaning, as well as the first paper that uses statistical model learning as a method to quantify meaning in the Semantic Web context. These insights are useful for the design of a new generation of Semantic Web tools that take such social meaning into account.
A Task-based Comparison of Linguistic and Semantic Document Retrieval Methods in the Medical Domain.
Shafahi, M., Hu, Q., Afsarmanesh, H., Huang, Z., ten Teije , A., & van Harmelen , F.
CEUR workshop proceedings, 1613. 2016.
link
bibtex
abstract
@article{39ec163cbc9a41078f10c194ebcd2f3e,
title = "A Task-based Comparison of Linguistic and Semantic Document Retrieval Methods in the Medical Domain",
abstract = "Text-based and semantics-based methods are both studied intensively as methods for document retrieval. In order to gain insight in the respective merits of these two approaches, we have performed a controlled experiment where we executed a real-life task using both textbased and semantics-based techniques. To maximise the lessons that we could draw about the two approaches, we have performed an experiment where we used the same task (searching papers from the scientific literature needed for updating a medical guideline), the same test-case (updating the 2004 Dutch national breast-cancer guideline), the same gold standard (the updated 2012 Dutch national breast-cancer guideline) and the same corpus (PubMed). We then performed this task using two different methods: retrieving papers based on keywords (text-based approach) and retrieving papers based on semantic annotations (semantics-based approach). Based on this experiment, we discuss the insights that we gained from this dual set of experiments.",
keywords = "Concept-based search, Document retrieval, Keyword search, Relation-based search, Semantic annotation",
author = "Mohammad Shafahi and Qing Hu and Hamideh Afsarmanesh and Z. Huang and {ten Teije}, A.C.M. and {van Harmelen}, F.A.H.",
year = "2016",
volume = "1613",
journal = "CEUR workshop proceedings",
issn = "1613-0073",
publisher = "CEUR Workshop Proceedings",
}
Text-based and semantics-based methods are both studied intensively as methods for document retrieval. In order to gain insight in the respective merits of these two approaches, we have performed a controlled experiment where we executed a real-life task using both textbased and semantics-based techniques. To maximise the lessons that we could draw about the two approaches, we have performed an experiment where we used the same task (searching papers from the scientific literature needed for updating a medical guideline), the same test-case (updating the 2004 Dutch national breast-cancer guideline), the same gold standard (the updated 2012 Dutch national breast-cancer guideline) and the same corpus (PubMed). We then performed this task using two different methods: retrieving papers based on keywords (text-based approach) and retrieving papers based on semantic annotations (semantics-based approach). Based on this experiment, we discuss the insights that we gained from this dual set of experiments.
.
Hu, Q., Huang, Z., ten Teije , A., van Harmelen , F., Marshall, M., & Dekker, A.
A topic-centric approach to detecting new evidences for evidence-based medical guidelines, pages 282–289. SciTePress, 2016.
link
bibtex
abstract
@inbook{91383e82af8f4f5c82634e58bfdbf443,
title = "A topic-centric approach to detecting new evidences for evidence-based medical guidelines",
abstract = "Evidence-based Medical guidelines are developed based on the best available evidence in biomedical science and clinical practice. Such evidence-based medical guidelines should be regularly updated, so that they can optimally serve medical practice by using the latest evidence from medical research. The usual approach to detect such new evidence is to use a set of terms from a guideline recommendation and to create queries for a biomedical search engine such as PubMed, with a ranking over a selected subset of terms to search for relevant new evidence. However, the terms that appear in a guideline recommendation do not always cover all of the information we need for the search, because the contextual information (e.g. time and location, user profile, topics) is usually missing in a guideline recommendation. Enhancing the search terms with contextual information would improve the quality of the search results. In this paper, we propose a topic-centric approach to detect new evidence for updating evidence-based medical guidelines as a context-aware method to improve the search. Our experiments show that this topic centric approach can find the goal evidence for 12 guideline statements out of 16 in our test set, compared with only 5 guideline statements that were found by using a non-topic centric approach.",
keywords = "Context-awareness, Evidence-based medical guidelines, Medical guideline update, Semantic distance, Topic-centric approach",
author = "Qing Hu and Zisheng Huang and {ten Teije}, Annette and {van Harmelen}, Frank and Marshall, {M. Scott} and Andre Dekker",
year = "2016",
pages = "282--289",
booktitle = "HEALTHINF 2016 - 9th International Conference on Health Informatics, Proceedings; Part of 9th International Joint Conference on Biomedical Engineering Systems and Technologies, BIOSTEC 2016",
publisher = "SciTePress",
}
Evidence-based Medical guidelines are developed based on the best available evidence in biomedical science and clinical practice. Such evidence-based medical guidelines should be regularly updated, so that they can optimally serve medical practice by using the latest evidence from medical research. The usual approach to detect such new evidence is to use a set of terms from a guideline recommendation and to create queries for a biomedical search engine such as PubMed, with a ranking over a selected subset of terms to search for relevant new evidence. However, the terms that appear in a guideline recommendation do not always cover all of the information we need for the search, because the contextual information (e.g. time and location, user profile, topics) is usually missing in a guideline recommendation. Enhancing the search terms with contextual information would improve the quality of the search results. In this paper, we propose a topic-centric approach to detect new evidence for updating evidence-based medical guidelines as a context-aware method to improve the search. Our experiments show that this topic centric approach can find the goal evidence for 12 guideline statements out of 16 in our test set, compared with only 5 guideline statements that were found by using a non-topic centric approach.
.
Pham, M., & Boncz, P.
Volume 9981 LNCS of Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). Exploiting emergent schemas to make RDF systems more efficient, pages 463–479. Springer/Verlag, 2016.
doi
link
bibtex
abstract
@inbook{489b9422588546ec9f2a8c9c806559fd,
title = "Exploiting emergent schemas to make RDF systems more efficient",
abstract = "We build on our earlier finding that more than 95% of the triples in actual RDF triple graphs have a remarkably tabular structure, whose schema does not necessarily follow from explicit metadata such as ontologies, but for which an RDF store can automatically derive by looking at the data using so-called “emergent schema” detection techniques. In this paper we investigate how computers and in particular RDF stores can take advantage from this emergent schema to more compactly store RDF data and more efficiently optimize and execute SPARQL queries. To this end, we contribute techniques for efficient emergent schema aware RDF storage and new query operator algorithms for emergent schema aware scans and joins. In all, these techniques allow RDF schema processors fully catch up with relational database techniques in terms of rich physical database design options and efficiency, without requiring a rigid upfront schema structure definition.",
author = "Pham, {Minh Duc} and Peter Boncz",
year = "2016",
doi = "10.1007/978-3-319-46523-4_28",
isbn = "9783319465227",
volume = "9981 LNCS",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
publisher = "Springer/Verlag",
pages = "463--479",
booktitle = "The Semantic Web - 15th International Semantic Web Conference, ISWC 2016, Proceedings",
}
We build on our earlier finding that more than 95% of the triples in actual RDF triple graphs have a remarkably tabular structure, whose schema does not necessarily follow from explicit metadata such as ontologies, but for which an RDF store can automatically derive by looking at the data using so-called “emergent schema” detection techniques. In this paper we investigate how computers and in particular RDF stores can take advantage from this emergent schema to more compactly store RDF data and more efficiently optimize and execute SPARQL queries. To this end, we contribute techniques for efficient emergent schema aware RDF storage and new query operator algorithms for emergent schema aware scans and joins. In all, these techniques allow RDF schema processors fully catch up with relational database techniques in terms of rich physical database design options and efficiency, without requiring a rigid upfront schema structure definition.
.
Carretta Zamborlini, V., Hoekstra, R., Da Silveira, M., Pruski, C., ten Teije , A., & van Harmelen , F.
Generalizing the Detection of Internal and External Interactions in Clinical Guidelines, pages 105–116. SciTePress, 2016.
doi
link
bibtex
abstract
@inbook{82b812259aed4390bb18653971bd6591,
title = "Generalizing the Detection of Internal and External Interactions in Clinical Guidelines",
abstract = "This paper presents a method for formally representing Computer-Interpretable Guidelines to deal with multimorbidity. Although some approaches for merging guidelines exist, improvements are still required for combining several sources of information and coping with possibly conflicting pieces of evidence coming from clinical studies. Our main contribution is twofold: (i) we provide general models and rules for representing guidelines that expresses evidence as causation beliefs; (ii) we introduce a mechanism to exploit external medical knowledge acquired from Linked Open Data (Drugbank, Sider, DIKB) to detect potential interactions between recommendations. We apply this framework to merge three guidelines (Osteoarthritis, Diabetes, and Hypertension) in order to illustrate the capability of this approach for detecting potential conflicts between guidelines and eventually propose alternatives.",
keywords = "Clinical guidelines, Knowledge representation, Ontologies, Semantic web",
author = "{Carretta Zamborlini}, Veruska and Rinke Hoekstra and {Da Silveira}, Marcos and Cedric Pruski and {ten Teije}, Annette and {van Harmelen}, Frank",
year = "2016",
doi = "10.5220/0005704101050116",
pages = "105--116",
booktitle = "HEALTHINF 2016 - 9th International Conference on Health Informatics, Proceedings; Part of 9th International Joint Conference on Biomedical Engineering Systems and Technologies, BIOSTEC 2016",
publisher = "SciTePress",
}
This paper presents a method for formally representing Computer-Interpretable Guidelines to deal with multimorbidity. Although some approaches for merging guidelines exist, improvements are still required for combining several sources of information and coping with possibly conflicting pieces of evidence coming from clinical studies. Our main contribution is twofold: (i) we provide general models and rules for representing guidelines that expresses evidence as causation beliefs; (ii) we introduce a mechanism to exploit external medical knowledge acquired from Linked Open Data (Drugbank, Sider, DIKB) to detect potential interactions between recommendations. We apply this framework to merge three guidelines (Osteoarthritis, Diabetes, and Hypertension) in order to illustrate the capability of this approach for detecting potential conflicts between guidelines and eventually propose alternatives.
grlc Makes GitHub Taste Like Linked Data APIs.
Meroño-Peñuela, A., & Hoekstra, R.
CEUR workshop proceedings, 1629. 2016.
doi
link
bibtex
abstract
@article{6dc9278699694bd4b43db86691dfa64c,
title = "grlc Makes GitHub Taste Like Linked Data APIs",
abstract = "Building Web APIs on top of SPARQL endpoints is becoming a common practice to enable universal access to the integration favorable dataspace of Linked Data. However, the Linked Data community cannot expect users to learn SPARQL to query this dataspace, and Web APIs are the most common way of enabling programmatic access to data on the Web. However, the implementation of Web APIs around Linked Data is often a tedious and repetitive process. Recent work speeds up thisLinked Data API construction by wrapping it around SPARQL queries, which carry out the API functionality under the hood. Inspired by this, in this paper we present grlc, a lightweight server that translates SPARQL queries curated in GitHub repositories to Linked Data APIs on the fly.",
keywords = "Git, GitHub, Linked data APIs, SPARQL",
author = "Albert Meroño-Peñuela and Rinke Hoekstra",
year = "2016",
doi = "10.1007/978-3-319-47602-5_48",
volume = "1629",
journal = "CEUR workshop proceedings",
issn = "1613-0073",
publisher = "CEUR Workshop Proceedings",
}
Building Web APIs on top of SPARQL endpoints is becoming a common practice to enable universal access to the integration favorable dataspace of Linked Data. However, the Linked Data community cannot expect users to learn SPARQL to query this dataspace, and Web APIs are the most common way of enabling programmatic access to data on the Web. However, the implementation of Web APIs around Linked Data is often a tedious and repetitive process. Recent work speeds up thisLinked Data API construction by wrapping it around SPARQL queries, which carry out the API functionality under the hood. Inspired by this, in this paper we present grlc, a lightweight server that translates SPARQL queries curated in GitHub repositories to Linked Data APIs on the fly.
.
Merono Penuela, A., & Hoekstra, R.
Volume 9989 LNCS of Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). grlc Makes GitHub Taste Like Linked Data APIs, pages 342–353. Springer/Verlag, 2016.
doi
link
bibtex
abstract
@inbook{44bee764c491494189957818fb36f123,
title = "grlc Makes GitHub Taste Like Linked Data APIs",
abstract = "Building Web APIs on top of SPARQL endpoints is becoming common practice. It enables universal access to the integration favorable data space of Linked Data. In the majority of use cases, users cannot be expected to learn SPARQL to query this data space. Web APIs are the most common way to enable programmatic access to data on the Web. However, the implementation of Web APIs around Linked Data is often a tedious and repetitive process. Recent work speeds up this Linked Data API construction by wrapping it around SPARQL queries, which carry out the API functionality under the hood. Inspired by this development, in this paper we present grlc, a lightweight server that takes SPARQL queries curated in GitHub repositories, and translates them to Linked Data APIs on the fly.",
keywords = "Git, GitHub, Linked data APIs, RESTFul, SPARQL",
author = "{Merono Penuela}, A. and R.J. Hoekstra",
year = "2016",
doi = "10.1007/978-3-319-47602-5_48",
isbn = "9783319476018",
volume = "9989 LNCS",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
publisher = "Springer/Verlag",
pages = "342--353",
booktitle = "The Semantic Web - ESWC 2016 Satellite Events, Revised Selected Papers",
}
Building Web APIs on top of SPARQL endpoints is becoming common practice. It enables universal access to the integration favorable data space of Linked Data. In the majority of use cases, users cannot be expected to learn SPARQL to query this data space. Web APIs are the most common way to enable programmatic access to data on the Web. However, the implementation of Web APIs around Linked Data is often a tedious and repetitive process. Recent work speeds up this Linked Data API construction by wrapping it around SPARQL queries, which carry out the API functionality under the hood. Inspired by this development, in this paper we present grlc, a lightweight server that takes SPARQL queries curated in GitHub repositories, and translates them to Linked Data APIs on the fly.
.
Leis, V., Gubichev, A., Mirchev, A., Boncz, P., Kemper, A., & Neumann, T.
Volume 9 . How Good Are Query Optimizers, Really?, pages 204–215. Association for Computing Machinery (ACM), 3 edition, 2016.
link
bibtex
abstract
@inbook{84bd5d6049614c52b3d233eeff1f1c76,
title = "How Good Are Query Optimizers, Really?",
abstract = "Finding a good join order is crucial for query performance. In this paper, we introduce the Join Order Benchmark (JOB) and experimentally revisit the main components in the classic query optimizer architecture using a complex, real-world data set and realistic multi-join queries. We investigate the quality of industrial-strength cardinality estimators and find that all estimators routinely produce large errors. We further show that while estimates are essential for finding a good join order, query performance is unsatisfactory if the query engine relies too heavily on these estimates. Using another set of experiments that measure the impact of the cost model, we find that it has much less influence on query performance than the cardinality estimates. Finally, we investigate plan enumeration techniques comparing exhaustive dynamic programming with heuristic algorithms and find that exhaustive enumeration improves performance despite the sub-optimal cardinality estimates.",
author = "Viktor Leis and Andrey Gubichev and Atanas Mirchev and Peter Boncz and Alfons Kemper and Thomas Neumann",
year = "2016",
volume = "9",
pages = "204--215",
booktitle = "Proceedings of the VLDB Endowment",
publisher = "Association for Computing Machinery (ACM)",
edition = "3",
}
Finding a good join order is crucial for query performance. In this paper, we introduce the Join Order Benchmark (JOB) and experimentally revisit the main components in the classic query optimizer architecture using a complex, real-world data set and realistic multi-join queries. We investigate the quality of industrial-strength cardinality estimators and find that all estimators routinely produce large errors. We further show that while estimates are essential for finding a good join order, query performance is unsatisfactory if the query engine relies too heavily on these estimates. Using another set of experiments that measure the impact of the cost model, we find that it has much less influence on query performance than the cardinality estimates. Finally, we investigate plan enumeration techniques comparing exhaustive dynamic programming with heuristic algorithms and find that exhaustive enumeration improves performance despite the sub-optimal cardinality estimates.
Inferring recommendation interactions in clinical guidelines.
Carretta Zamborlini, V., Hoekstra, R., Da Silveira, M., Pruski, C., ten Teije , A., & van Harmelen , F.
Semantic Web, 7(4): 421–446. 2016.
doi
link
bibtex
abstract
@article{4d178cf0055f4dd1acf70f17b8d01a25,
title = "Inferring recommendation interactions in clinical guidelines",
abstract = "The formal representation of clinical knowledge is still an open research topic. Classical representation languages for clinical guidelines are used to produce diagnostic and treatment plans. However, they have important limitations, e.g. when looking for ways to re-use, combine, and reason over existing clinical knowledge. These limitations are especially problematic in the context of multimorbidity; patients that suffer from multiple diseases. To overcome these limitations, this paper proposes a model for clinical guidelines (TMR4I) that allows the re-use and combination of knowledge from multiple guidelines. Semantic Web technology is applied to implement the model, allowing us to automatically infer interactions between recommendations, such as recommending the same drug more than once. It relies on an existing Linked Data set, DrugBank, for identifying drug-drug interactions. We evaluate the model by applying it to two realistic case studies on multimorbidity that combine guidelines for two (Duodenal Ulcer and Transient Ischemic Attack) and three diseases (Osteoarthritis, Hypertension and Diabetes) and compare the results with existing methods.",
keywords = "Clinical knowledge representation, OWL, SPARQL, SWRL, combining medical guidelines, multimorbidity, reasoning, rules",
author = "{Carretta Zamborlini}, Veruska and Rinke Hoekstra and {Da Silveira}, Marcos and Cedric Pruski and {ten Teije}, A.C.M. and {van Harmelen}, F.A.H.",
year = "2016",
doi = "10.3233/SW-150212",
volume = "7",
pages = "421--446",
journal = "Semantic Web",
issn = "1570-0844",
publisher = "IOS Press",
number = "4",
}
The formal representation of clinical knowledge is still an open research topic. Classical representation languages for clinical guidelines are used to produce diagnostic and treatment plans. However, they have important limitations, e.g. when looking for ways to re-use, combine, and reason over existing clinical knowledge. These limitations are especially problematic in the context of multimorbidity; patients that suffer from multiple diseases. To overcome these limitations, this paper proposes a model for clinical guidelines (TMR4I) that allows the re-use and combination of knowledge from multiple guidelines. Semantic Web technology is applied to implement the model, allowing us to automatically infer interactions between recommendations, such as recommending the same drug more than once. It relies on an existing Linked Data set, DrugBank, for identifying drug-drug interactions. We evaluate the model by applying it to two realistic case studies on multimorbidity that combine guidelines for two (Duodenal Ulcer and Transient Ischemic Attack) and three diseases (Osteoarthritis, Hypertension and Diabetes) and compare the results with existing methods.
Integrating Diachronous Conceptual Lexicons through Linked Open Data.
Maks, E., van Erp , M., Vossen, P., Hoekstra, R., & van der Sijs , N.
2016.
Proceedings title: DHBenelux Place of publication: Luxembourg
link
bibtex
@misc{60d8c039cc9e446685f5a19939709225,
title = "Integrating Diachronous Conceptual Lexicons through Linked Open Data",
author = "E. Maks and {van Erp}, M.G.J. and P.T.J.M. Vossen and R.J. Hoekstra and {van der Sijs}, N.",
note = "Proceedings title: DHBenelux Place of publication: Luxembourg",
year = "2016",
}
Interim report on the disambiguation results: RISIS Deliverable D25.1.
Schlobach, S., Idrissou, O., Hoekstra, R., Khalili, A., van Harmelen , F., & van den Besselaar , P.
Vrije Universiteit, 2016.
link
bibtex
@book{8575d1a8fc274801afefe62bcdff596a,
title = "Interim report on the disambiguation results: RISIS Deliverable D25.1",
author = "Stefan Schlobach and O.A.K. Idrissou and R.J. Hoekstra and A. Khalili and {van Harmelen}, Frank and {van den Besselaar}, P.A.A.",
year = "2016",
publisher = "Vrije Universiteit",
}
Linked Data Reactor: a Framework for Building Reactive Linked Data Applications.
Khalili, A.
CEUR workshop proceedings, 1615. 2016.
link
bibtex
abstract
@article{2405656e886645b996d943e2e03a1801,
title = "Linked Data Reactor: a Framework for Building Reactive Linked Data Applications",
abstract = "This paper presents Linked Data Reactor (LD-Reactor or LD-R) as a framework for developing exible and reusable User Interface components for Linked Data applications. LD-Reactor utilizes Facebook's ReactJS components, Flux architecture and Yahoo's Fluxible framework for isomorphic Web applications. It also exploits Semantic-UI framework for exible UI themes. LD-R aims to apply the idea of component-based application development into RDF data model hence enhancing current user interfaces to view, browse and edit Linked Data.",
author = "Ali Khalili",
year = "2016",
volume = "1615",
journal = "CEUR workshop proceedings",
issn = "1613-0073",
publisher = "CEUR Workshop Proceedings",
}
This paper presents Linked Data Reactor (LD-Reactor or LD-R) as a framework for developing exible and reusable User Interface components for Linked Data applications. LD-Reactor utilizes Facebook's ReactJS components, Flux architecture and Yahoo's Fluxible framework for isomorphic Web applications. It also exploits Semantic-UI framework for exible UI themes. LD-R aims to apply the idea of component-based application development into RDF data model hence enhancing current user interfaces to view, browse and edit Linked Data.
LOTUS: Adaptive Search for Big Linked Data.
Ilievski, F., Beek, W., van Erp , M., Rietveld, L., & Schlobach, K.
2016.
link
bibtex
@misc{9d1847e6571e4a07a23d8b2857aaed65,
title = "LOTUS: Adaptive Search for Big Linked Data",
author = "F. Ilievski and W.G.J. Beek and {van Erp}, M.G.J. and L.J. Rietveld and K.S. Schlobach",
year = "2016",
}
.
Ilievski, F., Beek, W., van Erp , M., Rietveld, L., & Schlobach, S.
of Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). LOTUS: Adaptive text search for big linked data, pages 470–485. Springer/Verlag, 2016.
doi
link
bibtex
abstract
@inbook{f61941f98374409c990fb5f25443b77d,
title = "LOTUS: Adaptive text search for big linked data",
abstract = "Finding relevant resources on the Semantic Web today is a dirty job: no centralized query service exists and the support for natural language access is limited. We present LOTUS: Linked Open Text Un- leaShed, a text-based entry point to a massive subset of today’s Linked Open Data Cloud. Recognizing the use case dependency of resource re- trieval, LOTUS provides an adaptive framework in which a set of match- ing and ranking algorithms are made available. Researchers and develop- ers are able to tune their own LOTUS index by choosing and combining the matching and ranking algorithms that suit their use case best. In this paper, we explain the LOTUS approach, its implementation and the functionality it provides. We demonstrate the ease with which LOTUS enables text-based resource retrieval at an unprecedented scale in con- crete and domain-specific scenarios. Finally, we provide evidence for the scalability of LOTUS with respect to the LOD Laundromat, the largest collection of easily accessible Linked Open Data currently available.",
keywords = "Findability, Scalable data management, Semantic search, Text indexing",
author = "F. Ilievski and Wouter Beek and {van Erp}, Marieke and Laurens Rietveld and Stefan Schlobach",
year = "2016",
doi = "10.1007/978-3-319-34129-3_29",
isbn = "9783319341286",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
publisher = "Springer/Verlag",
pages = "470--485",
booktitle = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
}
Finding relevant resources on the Semantic Web today is a dirty job: no centralized query service exists and the support for natural language access is limited. We present LOTUS: Linked Open Text Un- leaShed, a text-based entry point to a massive subset of today’s Linked Open Data Cloud. Recognizing the use case dependency of resource re- trieval, LOTUS provides an adaptive framework in which a set of match- ing and ranking algorithms are made available. Researchers and develop- ers are able to tune their own LOTUS index by choosing and combining the matching and ranking algorithms that suit their use case best. In this paper, we explain the LOTUS approach, its implementation and the functionality it provides. We demonstrate the ease with which LOTUS enables text-based resource retrieval at an unprecedented scale in con- crete and domain-specific scenarios. Finally, we provide evidence for the scalability of LOTUS with respect to the LOD Laundromat, the largest collection of easily accessible Linked Open Data currently available.
Phil@Scale: Computational methods within philosophy.
Van Wierst, P., Vrijenhoek, S., Schlobach, S., & Betti, A.
CEUR workshop proceedings, 1681. 2016.
link
bibtex
abstract
@article{9a629914bb37409db65a1698e87e8040,
title = "Phil@Scale: Computational methods within philosophy",
abstract = "In this paper we report the results of Phil@Scale, a project directed at the development of computational methods for (the history of) philosophy.1 In this project, philosophers and computer scientists together created SalVe, a tool that helps philosophers answering text-based questions. SalVe has been tested successfully on the Wissenschaftslehre (1837), an extensive work by the Bohemian polymath Bernard Bolzano (1781-1848). Bolzano was a philosopher, mathematician and theologian whose work has been of fundamental importance for the development of Western logic and the foundation of sciences such as mathematics and computer science. The testing of SalVe on the Wissenschaftslehre reveals that with respect to certain questions within philosophy valuable contributions are obtained by applying even rather simple, well-known computational techniques. We conclude that there is definitely a future for computational methods within text-based philosophical research. We explain how SalVe can be used within philosophical research that relies on textual sources. We will start out with an explanation of our aims in developing SalVe and give a short description of SalVe's functionalities, followed by a technical description of the tool. Then we will give a concrete example of how SalVe aids philosophical research. We conclude the paper with an evaluation of the potential of Digital Humanities tools for philosophy, and the challenges that face us if we wish to continue this development further.",
author = "{Van Wierst}, Pauline and Sanne Vrijenhoek and Stefan Schlobach and Arianna Betti",
year = "2016",
volume = "1681",
journal = "CEUR workshop proceedings",
issn = "1613-0073",
publisher = "CEUR Workshop Proceedings",
}
In this paper we report the results of Phil@Scale, a project directed at the development of computational methods for (the history of) philosophy.1 In this project, philosophers and computer scientists together created SalVe, a tool that helps philosophers answering text-based questions. SalVe has been tested successfully on the Wissenschaftslehre (1837), an extensive work by the Bohemian polymath Bernard Bolzano (1781-1848). Bolzano was a philosopher, mathematician and theologian whose work has been of fundamental importance for the development of Western logic and the foundation of sciences such as mathematics and computer science. The testing of SalVe on the Wissenschaftslehre reveals that with respect to certain questions within philosophy valuable contributions are obtained by applying even rather simple, well-known computational techniques. We conclude that there is definitely a future for computational methods within text-based philosophical research. We explain how SalVe can be used within philosophical research that relies on textual sources. We will start out with an explanation of our aims in developing SalVe and give a short description of SalVe's functionalities, followed by a technical description of the tool. Then we will give a concrete example of how SalVe aids philosophical research. We conclude the paper with an evaluation of the potential of Digital Humanities tools for philosophy, and the challenges that face us if we wish to continue this development further.
Publishing and Consuming Linked Data.
Rietveld, L.
2016.
Exacte Wetenschappen Naam instelling promotie: Vrije Universiteit Amsterdam Naam instelling onderzoek: Vrije Universiteit Amsterdam
link
bibtex
abstract
@misc{0bb6ee06d05a48a2b0000ac820655834,
title = "Publishing and Consuming Linked Data",
abstract = "13897",
author = "L.J. Rietveld",
note = "Exacte Wetenschappen Naam instelling promotie: Vrije Universiteit Amsterdam Naam instelling onderzoek: Vrije Universiteit Amsterdam",
year = "2016",
school = "Vrije Universiteit Amsterdam",
}
13897
Refining Statistical Data on the Web.
Merono Penuela, A.
2016.
Exacte Wetenschappen Naam instelling promotie: Vrije Universiteit Amsterdam Naam instelling onderzoek: Vrije Universiteit Amsterdam
link
bibtex
abstract
@misc{c61af5870a364375bc61f835f38c24de,
title = "Refining Statistical Data on the Web",
abstract = "13901",
author = "{Merono Penuela}, Albert",
note = "Exacte Wetenschappen Naam instelling promotie: Vrije Universiteit Amsterdam Naam instelling onderzoek: Vrije Universiteit Amsterdam",
year = "2016",
school = "Vrije Universiteit Amsterdam",
}
13901
SCRY: Extending SPARQL with custom data processing methods for the life sciences.
Stringer, B., Meroño-peñuela, A., Abeln, S., van Harmelen , F., & Heringa, J.
CEUR workshop proceedings, 1795: 1–10. 2016.
link
bibtex
@article{62ad2a6cd97a4be38502d5f84e54ec29,
title = "SCRY: Extending SPARQL with custom data processing methods for the life sciences",
keywords = "customization, data processing, rdf generation, SPARQL, extension",
author = "Bas Stringer and Albert Meroño-peñuela and Sanne Abeln and {van Harmelen}, Frank and Jaap Heringa",
year = "2016",
volume = "1795",
pages = "1--10",
journal = "CEUR workshop proceedings",
issn = "1613-0073",
publisher = "CEUR Workshop Proceedings",
}
Selected papers from the combined EKAW 2014 and Semantic Web journal track.
Schlobach, S., & Janowicz, K.
Semantic Web, 7(4): 333–334. 2016.
doi
link
bibtex
@article{48ce34ae6f064ad08f20a48bdfaaea18,
title = "Selected papers from the combined EKAW 2014 and Semantic Web journal track",
author = "Stefan Schlobach and Krzysztof Janowicz",
year = "2016",
doi = "10.3233/SW-160229",
volume = "7",
pages = "333--334",
journal = "Semantic Web",
issn = "1570-0844",
publisher = "IOS Press",
number = "4",
}
.
van den Besselaar , P., Khalili, A., Idrissou, O., Schlobach, K., & van Harmelen , F.
SMS: a linked open data infrastructure for science and innovation studies., pages 106–114. Rafols, I., editor(s). University Valencia, 2016.
link
bibtex
@inbook{6b8910eb91d746ed84e869023557eb2a,
title = "SMS: a linked open data infrastructure for science and innovation studies.",
author = "{van den Besselaar}, Peter and Ali Khalili and Oladele Idrissou and K.S. Schlobach and {van Harmelen}, F.A.H.",
year = "2016",
pages = "106--114",
editor = "Ismael Rafols",
booktitle = "Peripheries, Frontiers and Beyond; proceedings of the 21st STI Conference",
publisher = "University Valencia",
}
SWISH: An integrated semantic web notebook.
Beek, W., & Wielemaker, J.
CEUR workshop proceedings, 1690. 2016.
link
bibtex
@article{29b76753dd464f068489de2701391d51,
title = "SWISH: An integrated semantic web notebook",
author = "Wouter Beek and Jan Wielemaker",
year = "2016",
volume = "1690",
journal = "CEUR workshop proceedings",
issn = "1613-0073",
publisher = "CEUR Workshop Proceedings",
}
.
Beek, W., & Wielemaker, J.
SWISH: An Integrated Semantic Web Notebook. 2016.
link
bibtex
@inbook{888950bfc54943609de37cd94c2c9063,
title = "SWISH: An Integrated Semantic Web Notebook",
author = "Wouter Beek and Jan Wielemaker",
year = "2016",
booktitle = "Proceedings of the ISWC 2016 Posters Demonstrations Track co-located with 15th International Semantic Web Conference (ISWC 2016), Kobe, Japan, October 19, 2016.",
}
SWISH for prototyping clinical guideline interactions theory.
Carretta Zamborlini, V., Wielemaker, J., Da Silveira, M., Pruski, C., ten Teije , A., & van Harmelen , F.
CEUR workshop proceedings, 1795. 2016.
link
bibtex
abstract
@article{aeeb6a77c22d4ecebcef47cedec27955,
title = "SWISH for prototyping clinical guideline interactions theory",
abstract = "SWISH provides a general purpose collaborative infrastructure for applying Prolog reasoning over an RDF dataset together with features that facilitates prototyping Semantic Web applications. In this paper we report on the use of SWISH for efficiently developing a prototype for detection of clinical guideline interactions. These guidelines are a set of medical recommendations meant for supporting doctors on tackling a single disease. However, often guidelines need to be combined for treating patients that suffer from multiple diseases, and then a number of interactions can occur. The generic interaction rules are implemented in SWI-Prolog and the guideline RDF-data is enriched with clinical Linked Open Data (LOD) (e.g. Drugbank, Sider). We show the implementation of the proposed theory about interaction detection in a case-study on combining three guidelines. The experiment is interactively described using a SWISH notebook and the results are graphical visualised empowered by graphviz.",
keywords = "Clinical guideline interactions, Multimorbidity, Prolog, RDF, SWISH",
author = "{Carretta Zamborlini}, Veruska and Jan Wielemaker and {Da Silveira}, Marcos and Cedric Pruski and {ten Teije}, Annette and {van Harmelen}, F.A.H.",
year = "2016",
volume = "1795",
journal = "CEUR workshop proceedings",
issn = "1613-0073",
publisher = "CEUR Workshop Proceedings",
}
SWISH provides a general purpose collaborative infrastructure for applying Prolog reasoning over an RDF dataset together with features that facilitates prototyping Semantic Web applications. In this paper we report on the use of SWISH for efficiently developing a prototype for detection of clinical guideline interactions. These guidelines are a set of medical recommendations meant for supporting doctors on tackling a single disease. However, often guidelines need to be combined for treating patients that suffer from multiple diseases, and then a number of interactions can occur. The generic interaction rules are implemented in SWI-Prolog and the guideline RDF-data is enriched with clinical Linked Open Data (LOD) (e.g. Drugbank, Sider). We show the implementation of the proposed theory about interaction detection in a case-study on combining three guidelines. The experiment is interactively described using a SWISH notebook and the results are graphical visualised empowered by graphviz.
.
Valkering, O., de Boer , V., Lô, G., Blankendaal, R., & Schlobach, S.
Volume 10024 LNAI of Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). The semantic web in an SMS, pages 697–712. Springer/Verlag, 2016.
doi
link
bibtex
abstract
@inbook{4f2fea23a5e849ff801af292650d9fdf,
title = "The semantic web in an SMS",
abstract = "Many ICT applications and services, including those from the Semantic Web, rely on the Web for the exchange of data. This includes expensive server and network infrastructures. Most rural areas of developing countries are not reached by the Web and its possibilities, while at the same time the ability to share knowledge has been identified as a key enabler for development. To make widespread knowledge sharing possible in these rural areas, the notion of the Web has to be downscaled based on the specific low-resource infrastructure in place. In this paper, we introduce SPARQL over SMS, a solution for Web-like exchange of RDF data over cellular networks in which HTTP is substituted by SMS. We motivate and validate this through two use cases in West Africa. We present the design and implementation of the solution, along with a data compression method that combines generic compression strategies and strategies that use Semantic Web specific features to reduce the size of RDF before it is transferred over the low-bandwidth cellular network.",
author = "Onno Valkering and {de Boer}, Victor and Gossa Lô and Romy Blankendaal and Stefan Schlobach",
year = "2016",
doi = "10.1007/978-3-319-49004-5_45",
isbn = "9783319490038",
volume = "10024 LNAI",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
publisher = "Springer/Verlag",
pages = "697--712",
booktitle = "Knowledge Engineering and Knowledge Management - 20th International Conference, EKAW 2016, Proceedings",
}
Many ICT applications and services, including those from the Semantic Web, rely on the Web for the exchange of data. This includes expensive server and network infrastructures. Most rural areas of developing countries are not reached by the Web and its possibilities, while at the same time the ability to share knowledge has been identified as a key enabler for development. To make widespread knowledge sharing possible in these rural areas, the notion of the Web has to be downscaled based on the specific low-resource infrastructure in place. In this paper, we introduce SPARQL over SMS, a solution for Web-like exchange of RDF data over cellular networks in which HTTP is substituted by SMS. We motivate and validate this through two use cases in West Africa. We present the design and implementation of the solution, along with a data compression method that combines generic compression strategies and strategies that use Semantic Web specific features to reduce the size of RDF before it is transferred over the low-bandwidth cellular network.
.
Meroño-Peñuela, A., & Hoekstra, R.
Volume 9989 LNCS of Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). The song remains the same: Lossless conversion and streaming of MIDI to RDF and back, pages 194–199. Springer/Verlag, 2016.
doi
link
bibtex
abstract
@inbook{fa13d88f82b7462283a1cc9b8b232f44,
title = "The song remains the same: Lossless conversion and streaming of MIDI to RDF and back",
abstract = "In this demo, we explore the potential of RDF as a representation format for digital music. Digital music is broadly used today in many professional music production environments. For decades, MIDI (Musical Instrument Digital Interface) has been the standard for digital music exchange between musicians and devices, albeit not in a Web friendly way. We show the potential of expressing digital music as Linked Data, using our midi2rdf suite of tools to convert and stream digital music in MIDI format to RDF. The conversion allows for lossless round tripping: we can reconstruct a MIDI file identical to the original using its RDF representation. The streaming uses an existing, novel generative audio matching algorithm that we use to broadcast, with very low latency, RDF triples of MIDI events coming from arbitrary analog instruments.",
keywords = "Linked data, MIDI, Music streams, RDF",
author = "Albert Meroño-Peñuela and Rinke Hoekstra",
year = "2016",
doi = "10.1007/978-3-319-47602-5_38",
isbn = "9783319476018",
volume = "9989 LNCS",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
publisher = "Springer/Verlag",
pages = "194--199",
booktitle = "The Semantic Web - ESWC 2016 Satellite Events, Revised Selected Papers",
}
In this demo, we explore the potential of RDF as a representation format for digital music. Digital music is broadly used today in many professional music production environments. For decades, MIDI (Musical Instrument Digital Interface) has been the standard for digital music exchange between musicians and devices, albeit not in a Web friendly way. We show the potential of expressing digital music as Linked Data, using our midi2rdf suite of tools to convert and stream digital music in MIDI format to RDF. The conversion allows for lossless round tripping: we can reconstruct a MIDI file identical to the original using its RDF representation. The streaming uses an existing, novel generative audio matching algorithm that we use to broadcast, with very low latency, RDF triples of MIDI events coming from arbitrary analog instruments.
Towards an open data infrastructure for STI data. OECD Blue Sky Conference, Gent, September 2016.
van den Besselaar , P., Khalili, A., de Graaf , K., Idrissou, O., Loizou, A., Schlobach, K., & van Harmelen , F.
2016.
link
bibtex
@misc{7559caac92f74c2381a30ea519bbe559,
title = "Towards an open data infrastructure for STI data. OECD Blue Sky Conference, Gent, September 2016",
author = "{van den Besselaar}, Peter and Ali Khalili and {de Graaf}, {Klaas Andries} and Oladele Idrissou and A. Loizou and K.S. Schlobach and {van Harmelen}, F.A.H.",
year = "2016",
}
大数据时代的语义技术.
Huang, Z.
Journal of Digital Library Forum, 2016(10). 2016.
link
bibtex
@article{fd1c0df2db4246e19aabe172262616ca,
title = "大数据时代的语义技术",
author = "Zhisheng Huang",
year = "2016",
volume = "2016",
journal = "Journal of Digital Library Forum",
issn = "1673-2286",
number = "10",
}