\n \n \n
\n
\n\n \n \n \n \n \n \n Linked Open Data for Cultural Heritage.\n \n \n \n \n\n\n \n Alexiev, V.\n\n\n \n\n\n\n Ontotext webinar, 132 slides, September 2016.\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n \n pdf\n \n \n \n slides\n \n \n \n recording\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@Misc{Alexiev2016-CH-webinar,\n author = {Vladimir Alexiev},\n title = {{Linked Open Data for Cultural Heritage}},\n howpublished = {Ontotext webinar, 132 slides},\n month = sep,\n year = 2016,\n url = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20160929-webinar/index-full.html},\n url_PDF = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20160929-webinar/Linked%20Open%20Data%20for%20Cultural%20Heritage.pdf},\n url_Slides = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20160929-webinar/index.html},\n keywords = {AAT, Annotation, BTG, BTI, BTP, British Museum, Broader generic, Broader instantial, Broader partitive, CIDOC CRM, cultural heritage, EDM, ESE, Europeana, FRAD, FRBR, FRBR, FRBRoo, FRBRoo, FRSAD, Fundamental Concepts, Fundamental Relations, GLAM, Geonames, Getty, Getty Museum, ISNI, ISO 25964, LDBC, LOD, Metadata, Museum informatics, OAI, OAI PMH, OWLIM, Ontology, Ontotext GraphDB, Provenance, RDAinfo, RDF, ResearchSpace, SKOS, SKOS-XL, SPARQL, SPECTRUM, Schema, Seeing Standards, TGN, Taxonomy, Thesauri, ULAN, VIAF, Web Annotation, Wikidata, concept extraction, cultural heritage, dataset, endpoint, faceted search, food and drink, gazetteer, inference, knowledge base, knowledge-based system, multimedia annotation, ontology, open data, practical applications, reasoning, semantic application, semantic enrichment, semantic integration, semantic mapping, semantic repository, semantic representation, semantic search, semantic technology, text analysis, thesauri, virtual research environment, visualization, vocabularies},\n abstract = {The Internet, global digitization efforts, Europe's Digital Agenda, continuing investments in Europeana, the Digital Public Library of America and many other initiatives, have made millions upon millions of digitized cultural artifacts available on the net. We need to make sense of all this information: aggregate it, integrate it, provide cross-collection search, find links between entities and artefacts, build narratives, analyze data, support the scientific discourse, engage users… From ancient maps to bibliographic records, to paintings, to coins and hoards, to paleographic analysis, to prosopography factoids... everything is becoming more and more connected. A host of ontologies and metadata standards exist in the Cultural Heritage (CH) domain: CIDOC CRM, TEI5, LIDO, SPECTRUM, VRA Core, MPEG7, DC, ESE and EDM, OAI ORE and PMH, IIIIF, ResourceSync... the list goes on and on. How many of the standards listed in Seeing Standards: A Visualization of the Metadata Universe (by Jenn Riley, Associate Dean for Digital Initiatives at McGill University Library) apply to your work? A number of established thesauri and gazetteers exist, and some of them are interconnected: DBPedia; Wikidata, VIAF, FAST, ULAN; GeoNames, Pleiades, TGN; LCSH, AAT, IconClass, Joconde, SVCN, Wordnet, etc etc. The diagram below (by Michiel Hildebrand) shows a small part of this upcoming universe of CH data. How to use them in every-day collection management, cataloging, documentation and research? How to expose your institution's collections and other data to allow interlinking? Digital Humanities (DH) has emerged as a new and promising scientific discipline, with universities like Kings College London establishing new departments devoted to it. As Jeffrey Schnapp writes in the Digital Humanities manifesto 2.0 "Digital Humanities embraces and harnesses the expanded, global nature of today’s research communities as one of the great inter-disciplinary/post-disciplinary opportunities of our time. It dreams of models of knowledge production and reproduction that leverage the increasingly distributed nature of expertise and knowledge and transform this reality into occasions for scholarly innovation, disciplinary cross-fertilization, and the democratization of knowledge". In his keynote address at MCN 2014 Beyond Borders: The Humanities in the Digital Age, James Cuno (President and CEO of the J. Paul Getty Trust) emphasizes the role of modernizing Humanities and the value of Linked Data in cultural heritage informatics. The question also is how to preserve the role of libraries, museums and other Cultural Heritage institutions as centers of wisdom and culture into the new millennium? Aren't Google, Wikipedia, Facebook, Twitter and smart-phone apps becoming the new centers of research and culture (or at least popular culture)? We believe the answers to many of these questions lie with Semantic Technology and Linked Data. They enable large-scale Digital Humanities research, collaboration and aggregation; and technological renewal of CH institutions. The Rosetta Stone was key to the deciphering of Egyptian hieroglyphs, by providing parallel text in three scripts: Ancient Egyptian, Demotic and Ancient Greek. Today semantic technologies play a similar role, allowing the Digital Humanist to make connections between (and make sense of) the multitude of digitized cultural artifacts available on the net. An upsurge of interest in semantic technology has swept the CH and DH communities. Meetups and summits, conferences and un-conferences, residences and hackathons are taking place every week. CH institutions are collaborating actively. An active Linked Open Data for Libraries, Archives and Museums (LODLAM) community has emerged, and the #LODLAM twitter hashtag sees active communication. Established institutions create branches that sound like web startups or Wikipedia offsprings (e.g. British Library Labs; Smithsonian Web-strategy and Smithsonian Commons; UK National Archives department of Web Continuity). The Galleries, Libraries, Archives and Museums (GLAM) sector deals with complex and varied data. Integrating that data, especially across institutions, has always been a challenge. On the other hand, the value of linked data is especially high in this sector, since culture by its very nature is cross-border and interlinked. In this webinar we'll present interesting LODLAM projects, datasets and ontologies, as well as Ontotext's experience in this domain.},\n url_recording= {https://ontotext.com/knowledgehub/webinars/build-narratives-connect-artifacts-cultural-heritage/},\n}\n\n
\n
\n\n\n
\n The Internet, global digitization efforts, Europe's Digital Agenda, continuing investments in Europeana, the Digital Public Library of America and many other initiatives, have made millions upon millions of digitized cultural artifacts available on the net. We need to make sense of all this information: aggregate it, integrate it, provide cross-collection search, find links between entities and artefacts, build narratives, analyze data, support the scientific discourse, engage users… From ancient maps to bibliographic records, to paintings, to coins and hoards, to paleographic analysis, to prosopography factoids... everything is becoming more and more connected. A host of ontologies and metadata standards exist in the Cultural Heritage (CH) domain: CIDOC CRM, TEI5, LIDO, SPECTRUM, VRA Core, MPEG7, DC, ESE and EDM, OAI ORE and PMH, IIIIF, ResourceSync... the list goes on and on. How many of the standards listed in Seeing Standards: A Visualization of the Metadata Universe (by Jenn Riley, Associate Dean for Digital Initiatives at McGill University Library) apply to your work? A number of established thesauri and gazetteers exist, and some of them are interconnected: DBPedia; Wikidata, VIAF, FAST, ULAN; GeoNames, Pleiades, TGN; LCSH, AAT, IconClass, Joconde, SVCN, Wordnet, etc etc. The diagram below (by Michiel Hildebrand) shows a small part of this upcoming universe of CH data. How to use them in every-day collection management, cataloging, documentation and research? How to expose your institution's collections and other data to allow interlinking? Digital Humanities (DH) has emerged as a new and promising scientific discipline, with universities like Kings College London establishing new departments devoted to it. As Jeffrey Schnapp writes in the Digital Humanities manifesto 2.0 \"Digital Humanities embraces and harnesses the expanded, global nature of today’s research communities as one of the great inter-disciplinary/post-disciplinary opportunities of our time. It dreams of models of knowledge production and reproduction that leverage the increasingly distributed nature of expertise and knowledge and transform this reality into occasions for scholarly innovation, disciplinary cross-fertilization, and the democratization of knowledge\". In his keynote address at MCN 2014 Beyond Borders: The Humanities in the Digital Age, James Cuno (President and CEO of the J. Paul Getty Trust) emphasizes the role of modernizing Humanities and the value of Linked Data in cultural heritage informatics. The question also is how to preserve the role of libraries, museums and other Cultural Heritage institutions as centers of wisdom and culture into the new millennium? Aren't Google, Wikipedia, Facebook, Twitter and smart-phone apps becoming the new centers of research and culture (or at least popular culture)? We believe the answers to many of these questions lie with Semantic Technology and Linked Data. They enable large-scale Digital Humanities research, collaboration and aggregation; and technological renewal of CH institutions. The Rosetta Stone was key to the deciphering of Egyptian hieroglyphs, by providing parallel text in three scripts: Ancient Egyptian, Demotic and Ancient Greek. Today semantic technologies play a similar role, allowing the Digital Humanist to make connections between (and make sense of) the multitude of digitized cultural artifacts available on the net. An upsurge of interest in semantic technology has swept the CH and DH communities. Meetups and summits, conferences and un-conferences, residences and hackathons are taking place every week. CH institutions are collaborating actively. An active Linked Open Data for Libraries, Archives and Museums (LODLAM) community has emerged, and the #LODLAM twitter hashtag sees active communication. Established institutions create branches that sound like web startups or Wikipedia offsprings (e.g. British Library Labs; Smithsonian Web-strategy and Smithsonian Commons; UK National Archives department of Web Continuity). The Galleries, Libraries, Archives and Museums (GLAM) sector deals with complex and varied data. Integrating that data, especially across institutions, has always been a challenge. On the other hand, the value of linked data is especially high in this sector, since culture by its very nature is cross-border and interlinked. In this webinar we'll present interesting LODLAM projects, datasets and ontologies, as well as Ontotext's experience in this domain.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Using DBPedia in Europeana Food and Drink.\n \n \n \n \n\n\n \n Alexiev, V.\n\n\n \n\n\n\n presentation, February 2016.\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n \n pdf\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@Misc{Alexiev2016-EFD-DBpedia,\n author = {Vladimir Alexiev},\n title = {{Using DBPedia in Europeana Food and Drink}},\n howpublished = {presentation},\n month = feb,\n year = 2016,\n url = {https://drive.google.com/file/d/0B7je1jgVmCgIZzNiWmdqTGpDa28/view},\n url_PDF = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20160212-Using-DBPedia-in-Europeana-Food-and-Drink.pdf},\n keywords = {Europeana, cultural heritage, food and drink, DBpedia, Geonames, semantic application, faceted search, semantic search},\n address = {DBpedia Meeting, The Hague, Netherlands},\n abstract = {The Europeana Food and Drink project collects cultural heritage objects for and develops applications related to Food and Drink heritage. As part of the project, Ontotext developed a FD Classification based on Wikipedia/DBpedia Categories, a semantic enrichment service that annotates each CHO with FD Topics and Places, and a semantic application (https://efd.ontotext.com/app) that implements hierarchical semantic facets and semantic search for these facets. We'll also be packaging the enrichment as a service for others to use in a crowdsourced annotation application. We will explain how we used Categories to build a domain-specific gazetteer, used external datasets (eg UMBEL domains and DBTax types), correlated DBpedia places to Geonames to use the place hierarchy, and the workings of the semantic application},\n}\n\n
\n
\n\n\n
\n The Europeana Food and Drink project collects cultural heritage objects for and develops applications related to Food and Drink heritage. As part of the project, Ontotext developed a FD Classification based on Wikipedia/DBpedia Categories, a semantic enrichment service that annotates each CHO with FD Topics and Places, and a semantic application (https://efd.ontotext.com/app) that implements hierarchical semantic facets and semantic search for these facets. We'll also be packaging the enrichment as a service for others to use in a crowdsourced annotation application. We will explain how we used Categories to build a domain-specific gazetteer, used external datasets (eg UMBEL domains and DBTax types), correlated DBpedia places to Geonames to use the place hierarchy, and the workings of the semantic application\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Europeana Food and Drink Semantic Demonstrator Extended.\n \n \n \n \n\n\n \n Alexiev, V.; Tagarev, A.; and Tolosi, L.\n\n\n \n\n\n\n Technical Report D3.20d, Europeana Food and Drink project, July 2016.\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@TechReport{Alexiev2016-EFD-semapp-ext,\n author = {Vladimir Alexiev and Andrey Tagarev and Laura Tolosi},\n title = {{Europeana Food and Drink Semantic Demonstrator Extended}},\n institution = {Europeana Food and Drink project},\n year = 2016,\n type = {Deliverable},\n number = {D3.20d},\n month = jul,\n url = {https://rawgit2.com/VladimirAlexiev/my/master/pubs/Europeana-Food-and-Drink-Semantic-Demonstrator-Extended-(D3.20d).pdf},\n keywords = {Europeana, cultural heritage, food and drink, semantic application, semantic search, faceted search, semantic enrichment},\n abstract = {Describes the additional development on the EFD Semantic Demonstrator performed after the official D3.20 deliverable (M22). It describes work performed between 31 October 2015 and 20 July 2016 (M31), the achieved results, the created data and enrichments, and the extended application functionality.},\n}\n\n
\n
\n\n\n
\n Describes the additional development on the EFD Semantic Demonstrator performed after the official D3.20 deliverable (M22). It describes work performed between 31 October 2015 and 20 July 2016 (M31), the achieved results, the created data and enrichments, and the extended application functionality.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Meet the Europeana Members Council: Vladimir Alexiev.\n \n \n \n \n\n\n \n Alexiev, V.\n\n\n \n\n\n\n blog post, March 2016.\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@Misc{Alexiev2016-EuropeanaMC-blog,\n author = {Vladimir Alexiev},\n title = {{Meet the Europeana Members Council: Vladimir Alexiev}},\n howpublished = {blog post},\n month = mar,\n year = 2016,\n url = {https://pro.europeana.eu/blogpost/meet-the-members-council-vladimir-alexiev},\n keywords = {cultural heritage, Europeana, EHRI, ResearchSpace, data quality, semantic enrichment},\n abstract = {Describes the work of Ontotext and in particular Vladimir Alexiev in applying semantic technologies to cultural heritage},\n}\n\n
\n
\n\n\n
\n Describes the work of Ontotext and in particular Vladimir Alexiev in applying semantic technologies to cultural heritage\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Getty Vocabularies: LOD Sample Queries.\n \n \n \n \n\n\n \n Alexiev, V.\n\n\n \n\n\n\n Getty Research Institute, 3.3 edition, May 2016.\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@Manual{Alexiev2016-GVP-LOD-queries,\n title = {{Getty Vocabularies: LOD Sample Queries}},\n author = {Vladimir Alexiev},\n organization = {Getty Research Institute},\n edition = {3.3},\n month = may,\n year = 2016,\n url = {https://vocab.getty.edu/doc/queries/},\n abstract = {We provide 120 sample queries for the Getty Vocabularies LOD that should allow you to learn to query the data effectively. We include searching for data, getting all data of a subject, all labels and their attributes, full-text search, getting an ordered hierarchy, charts, etc. The queries are organized in sections: general, TGN-specific, ULAN-specific, Language queries, Counting and descriptive info, Exploring the ontology},\n keywords = {Getty, GVP, vocabularies, thesauri, AAT, TGN, ULAN, SPARQL, ontology, SKOS, SKOS-XL, ISO 25964},\n}\n\n
\n
\n\n\n
\n We provide 120 sample queries for the Getty Vocabularies LOD that should allow you to learn to query the data effectively. We include searching for data, getting all data of a subject, all labels and their attributes, full-text search, getting an ordered hierarchy, charts, etc. The queries are organized in sections: general, TGN-specific, ULAN-specific, Language queries, Counting and descriptive info, Exploring the ontology\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n How Not to Do Linked Data.\n \n \n \n \n\n\n \n Alexiev, V.\n\n\n \n\n\n\n Github gist, December 2016.\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@Misc{Alexiev2016-How-not-to-do-LOD,\n author = {Vladimir Alexiev},\n title = {{How Not to Do Linked Data}},\n howpublished = {Github gist},\n month = dec,\n year = 2016,\n url = {https://gist.github.com/VladimirAlexiev/090d5e54a525d57acb9b366121e77573},\n keywords = {cultural heritage, RDF, LODLAM, CIDOC CRM, mapping, review, data quality},\n abstract = {I review the data quality of the LOD publication of a national cultural heritage institution, and show examples of bad practices},\n}\n\n
\n
\n\n\n
\n I review the data quality of the LOD publication of a national cultural heritage institution, and show examples of bad practices\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Multisensor RDF Application Profile.\n \n \n \n \n\n\n \n Alexiev, V.\n\n\n \n\n\n\n Technical Report Multisensor Project, Ontotext Corp, October 2016.\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n \n source\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@TechReport{Alexiev2016-Multisensor-profile,\n author = {Vladimir Alexiev},\n title = {{Multisensor RDF Application Profile}},\n institution = {Multisensor Project, Ontotext Corp},\n year = 2016,\n month = oct,\n abstract = {The Multisensor project analyzes and extracts data from mass- and social media documents (so-called SIMMOs), including text, images and video, speech recognition and translationn, across several languages. It also handles social network data, statistical data, etc. Early on the project made the decision that all data exchanged between project partners (between modules inside and outside the processing pipeline) will be in RDF JSONLD format. The final data is stored in a semantic repository and is used by various User Interface components for end-user interaction. This final data forms a corpus of semantic data over SIMMOs and is an important outcome of the project. The flexibility of the semantic web model has allowed us to accommodate a huge variety of data in the same extensible model. We use a number of ontologies for representing that data: NIF and OLIA for linguistic info, ITSRDF for NER, DBpedia and Babelnet for entities and concepts, MARL for sentiment, OA for image and cross-article annotations, W3C CUBE for statistical indicators, etc. In addition to applying existing ontologies, we extended them by the Multisensor ontology, and introduced some innovations like embedding FrameNet in NIF. The documentation of this data has been an important ongoing task. It is even more important towards the end of the project, in order to enable the efficient use of MS data by external consumers. This document describes the different RDF patterns used by Multisensor, and how the data fits together. Thus it represents an "RDF Application Profile" for Multisensor. We use an example-based approach, rather than the more formal and labourious approach being standardized by the W3C RDF Shapes working group (still in development). We cover the following areas: 1. Linguistic Linked Data in NLP Interchange Format (NIF), including Part of Speech (POS), dependency parsing, sentiment, Named Entity Recognition (NER), etc. 2. Speech recognition, translation. 3. Multimedia binding and image annotation. 4. Statistical indicators and similar data. 5. Social network popularity and influence, etc.},\n url = {https://rawgit2.com/VladimirAlexiev/multisensor/master/index.html},\n url_Source = {https://github.com/VladimirAlexiev/multisensor},\n keywords = {Multisensor, CUBE, NLP, NLP2RDF, NIF, OLIA, ITSRDF, NERD, MARL, BabelNet, FrameNet, WordNet},\n}\n\n
\n
\n\n\n
\n The Multisensor project analyzes and extracts data from mass- and social media documents (so-called SIMMOs), including text, images and video, speech recognition and translationn, across several languages. It also handles social network data, statistical data, etc. Early on the project made the decision that all data exchanged between project partners (between modules inside and outside the processing pipeline) will be in RDF JSONLD format. The final data is stored in a semantic repository and is used by various User Interface components for end-user interaction. This final data forms a corpus of semantic data over SIMMOs and is an important outcome of the project. The flexibility of the semantic web model has allowed us to accommodate a huge variety of data in the same extensible model. We use a number of ontologies for representing that data: NIF and OLIA for linguistic info, ITSRDF for NER, DBpedia and Babelnet for entities and concepts, MARL for sentiment, OA for image and cross-article annotations, W3C CUBE for statistical indicators, etc. In addition to applying existing ontologies, we extended them by the Multisensor ontology, and introduced some innovations like embedding FrameNet in NIF. The documentation of this data has been an important ongoing task. It is even more important towards the end of the project, in order to enable the efficient use of MS data by external consumers. This document describes the different RDF patterns used by Multisensor, and how the data fits together. Thus it represents an \"RDF Application Profile\" for Multisensor. We use an example-based approach, rather than the more formal and labourious approach being standardized by the W3C RDF Shapes working group (still in development). We cover the following areas: 1. Linguistic Linked Data in NLP Interchange Format (NIF), including Part of Speech (POS), dependency parsing, sentiment, Named Entity Recognition (NER), etc. 2. Speech recognition, translation. 3. Multimedia binding and image annotation. 4. Statistical indicators and similar data. 5. Social network popularity and influence, etc.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n How to find Open Data and Ontologies in Linguistics/NLP and Cultural Heritage.\n \n \n \n \n\n\n \n Alexiev, V.\n\n\n \n\n\n\n presentation, March 2016.\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n \n slides\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@Misc{Alexiev2016-OpenData,\n author = {Vladimir Alexiev},\n title = {{How to find Open Data and Ontologies in Linguistics/NLP and Cultural Heritage}},\n howpublished = {presentation},\n month = mar,\n year = 2016,\n url = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20160329-OpenData-and-Ontologies/index-full.html},\n url_Slides = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20160329-OpenData-and-Ontologies/index.html},\n keywords = {open data, ontology, linguistiscs, NLP, cultural heritage},\n address = {4th Open Data & Linked Data meetup, Sofia, Bulgaria},\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Multisensor Linked Open Data.\n \n \n \n \n\n\n \n Alexiev, V.\n\n\n \n\n\n\n presentation, September 2016.\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@Misc{Alexiev2016-dbpedia-multisensor,\n author = {Vladimir Alexiev},\n title = {{Multisensor Linked Open Data}},\n month = sep,\n year = 2016,\n url = {https://rawgit2.com/VladimirAlexiev/multisensor/master/20160915-Multisensor-LOD/index.html},\n keywords = {Multisensor, CUBE, NLP, NLP2RDF, NIF, OLIA, ITSRDF, NERD, MARL, BabelNet, FrameNet, WordNet},\n booktitle = {{DBpedia Meeting}},\n howpublished = {presentation},\n address = {Leipzig, Germany},\n abstract = {The FP7 Multisensor project analyzes and extracts data from mass- and social media documents, including text, images and video, across several languages. It uses a number of ontologies for representing that data: NIF and OLIA for linguistic info, ITSRDF for NER, DBpedia and Babelnet for entities and concepts, MARL for sentiment, OA for image and cross-article annotations, etc. We'll present how all these ontologies fit together, and some innovations like embedding FrameNet in NIF.},\n}\n\n
\n
\n\n\n
\n The FP7 Multisensor project analyzes and extracts data from mass- and social media documents, including text, images and video, across several languages. It uses a number of ontologies for representing that data: NIF and OLIA for linguistic info, ITSRDF for NER, DBpedia and Babelnet for entities and concepts, MARL for sentiment, OA for image and cross-article annotations, etc. We'll present how all these ontologies fit together, and some innovations like embedding FrameNet in NIF.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Making True RDF Diagrams with rdfpuml.\n \n \n \n \n\n\n \n Alexiev, V.\n\n\n \n\n\n\n presentation, March 2016.\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n \n slides\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@Misc{Alexiev2016-rdfpuml,\n author = {Vladimir Alexiev},\n title = {{Making True RDF Diagrams with rdfpuml}},\n howpublished = {presentation},\n month = mar,\n year = 2016,\n url = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20160514-rdfpuml/index-full.html},\n url_Slides = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20160514-rdfpuml/index.html},\n keywords = {RDF, visualization, PlantUML, cultural heritage, NLP, NIF, EHRI},\n abstract = {RDF is a graph data model, thus often the best way to understand RDF data schemas (ontologies, application profiles, RDF shapes) is with a diagram. We describe a tool (rdfpuml) that makes true diagrams from Turtle examples using PlantUML and GraphViz. Diagram readability is of prime concern, and rdfpuml introduces a few diagram control mechanisms using triples in the puml: namespace. We give examples from Getty CONA (Mappings of museum data to CIDOC CRM), Multisensor (NLP2RDF/NIF, FrameNet), EHRI (Holocaust Research into Jewish social networks), Duraspace (Portland Common Data Model for holding metadata in institutional repositories)},\n}\n\n
\n
\n\n\n
\n RDF is a graph data model, thus often the best way to understand RDF data schemas (ontologies, application profiles, RDF shapes) is with a diagram. We describe a tool (rdfpuml) that makes true diagrams from Turtle examples using PlantUML and GraphViz. Diagram readability is of prime concern, and rdfpuml introduces a few diagram control mechanisms using triples in the puml: namespace. We give examples from Getty CONA (Mappings of museum data to CIDOC CRM), Multisensor (NLP2RDF/NIF, FrameNet), EHRI (Holocaust Research into Jewish social networks), Duraspace (Portland Common Data Model for holding metadata in institutional repositories)\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n RDF by Example: rdfpuml for True RDF Diagrams, rdf2rml for R2RML Generation.\n \n \n \n \n\n\n \n Alexiev, V.\n\n\n \n\n\n\n In
Semantic Web in Libraries 2016 (SWIB 2016), Bonn, Germany, November 2016. \n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n \n slides\n \n \n \n video\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@InProceedings{Alexiev2016-rdfpuml-rdf2rml,\n author = {Vladimir Alexiev},\n title = {{RDF by Example: rdfpuml for True RDF Diagrams, rdf2rml for R2RML Generation}},\n booktitle = {{Semantic Web in Libraries 2016 (SWIB 2016)}},\n year = 2016,\n month = nov,\n address = {Bonn, Germany},\n url = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20161128-rdfpuml-rdf2rml/index-full.html},\n url_Slides = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20161128-rdfpuml-rdf2rml/index.html},\n url_Video = {https://youtu.be/4WoYlaGF6DE},\n keywords = {RDF, visualization, PlantUML, cultural heritage, NLP, NIF, EHRI, R2RML, generation, model-driven, RDF by Example, rdfpuml, rdf2rml},\n abstract = {RDF is a graph data model, so the best way to understand RDF data schemas (ontologies, application profiles, RDF shapes) is with a diagram. Many RDF visualization tools exist, but they either focus on large graphs (where the details are not easily visible), or the visualization results are not satisfactory, or manual tweaking of the diagrams is required. We describe a tool *rdfpuml* that makes true diagrams directly from Turtle examples using PlantUML and GraphViz. Diagram readability is of prime concern, and rdfpuml introduces various diagram control mechanisms using triples in the puml: namespace. Special attention is paid to inlining and visualizing various Reification mechanisms (described with PRV). We give examples from Getty CONA, Getty Museum, AAC (mappings of museum data to CIDOC CRM), Multisensor (NIF and FrameNet), EHRI (Holocaust Research into Jewish social networks), Duraspace (Portland Common Data Model for holding metadata in institutional repositories), Video annotation. If the example instances include SQL queries and embedded field names, they can describe a mapping precisely. Another tool *rdf2rdb* generates R2RML transformations from such examples, saving about 15x in complexity.},\n}\n% Future work: extend RDF by Example to describe RDF Shapes; extend rdf2rml to generate RML instead of only R2RML, i.e. handle XML and JSON data sources\n% https://docs.stardog.com/#_stardog_mapping_syntax is similar: shortcut syntax of R2RML that displays examples\n\n
\n
\n\n\n
\n RDF is a graph data model, so the best way to understand RDF data schemas (ontologies, application profiles, RDF shapes) is with a diagram. Many RDF visualization tools exist, but they either focus on large graphs (where the details are not easily visible), or the visualization results are not satisfactory, or manual tweaking of the diagrams is required. We describe a tool *rdfpuml* that makes true diagrams directly from Turtle examples using PlantUML and GraphViz. Diagram readability is of prime concern, and rdfpuml introduces various diagram control mechanisms using triples in the puml: namespace. Special attention is paid to inlining and visualizing various Reification mechanisms (described with PRV). We give examples from Getty CONA, Getty Museum, AAC (mappings of museum data to CIDOC CRM), Multisensor (NIF and FrameNet), EHRI (Holocaust Research into Jewish social networks), Duraspace (Portland Common Data Model for holding metadata in institutional repositories), Video annotation. If the example instances include SQL queries and embedded field names, they can describe a mapping precisely. Another tool *rdf2rdb* generates R2RML transformations from such examples, saving about 15x in complexity.\n
\n\n\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n The health care and life sciences community profile for dataset descriptions.\n \n \n \n \n\n\n \n Dumontier, M.; Gray, A. J. G.; Marshall, M. S.; Alexiev, V.; and others\n\n\n \n\n\n\n
PeerJ, 4: e2331. August 2016.\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@Article{HCLS-paper,\n author = {Michel Dumontier and Alasdair J. G. Gray and M. Scott Marshall and Vladimir Alexiev and others},\n title = {{The health care and life sciences community profile for dataset descriptions}},\n journal = {{PeerJ}},\n year = 2016,\n volume = 4,\n pages = {e2331},\n month = aug,\n url = {https://peerj.com/articles/2331/},\n keywords = {Data profiling, Dataset descriptions, Metadata, Provenance, FAIR data, HCLS, dataset, VOID, ontology, Bioinformatics, Taxonomy},\n issn = {2167-8359},\n doi = {10.7717/peerj.2331},\n abstract = {Access to consistent, high-quality metadata is critical to finding, understanding, and reusing scientific data. However, while there are many relevant vocabularies for the annotation of a dataset, none sufficiently captures all the necessary metadata. This prevents uniform indexing and querying of dataset repositories. Towards providing a practical guide for producing a high quality description of biomedical datasets, the W3C Semantic Web for Health Care and the Life Sciences Interest Group (HCLSIG) identified RDF vocabularies that could be used to specify common metadata elements and their value sets. The resulting guideline covers elements of description, identification, attribution, versioning, provenance, and content summarization. This guideline reuses existing vocabularies, and is intended to meet key functional requirements including indexing, discovery, exchange, query, and retrieval of datasets. The resulting metadata profile is generic and could be used by other domains with an interest in providing machine readable descriptions of versioned datasets.},\n}\n\n
\n
\n\n\n
\n Access to consistent, high-quality metadata is critical to finding, understanding, and reusing scientific data. However, while there are many relevant vocabularies for the annotation of a dataset, none sufficiently captures all the necessary metadata. This prevents uniform indexing and querying of dataset repositories. Towards providing a practical guide for producing a high quality description of biomedical datasets, the W3C Semantic Web for Health Care and the Life Sciences Interest Group (HCLSIG) identified RDF vocabularies that could be used to specify common metadata elements and their value sets. The resulting guideline covers elements of description, identification, attribution, versioning, provenance, and content summarization. This guideline reuses existing vocabularies, and is intended to meet key functional requirements including indexing, discovery, exchange, query, and retrieval of datasets. The resulting metadata profile is generic and could be used by other domains with an interest in providing machine readable descriptions of versioned datasets.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Semantic Integration of Web Data for International Investment Decision Support.\n \n \n \n \n\n\n \n Simeonov, B.; Alexiev, V.; Liparas, D.; Puigbo, M.; Vrochidis, S.; Jamin, E.; and Kompatsiaris, I.\n\n\n \n\n\n\n In
3rd International Conference on Internet Science (INSCI 2016), Florence, Italy, September 2016. \n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n \n preprint\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@InProceedings{INSCI2016-Multisensor,\n author = {Boyan Simeonov and Vladimir Alexiev and Dimitris Liparas and Marti Puigbo and Stefanos Vrochidis and Emmanuel Jamin and Ioannis Kompatsiaris},\n title = {{Semantic Integration of Web Data for International Investment Decision Support}},\n booktitle = {{3rd International Conference on Internet Science (INSCI 2016)}},\n year = 2016,\n month = sep,\n address = {Florence, Italy},\n url = {https://zenodo.org/record/571202},\n url_Preprint = {https://rawgit2.com/VladimirAlexiev/my/master/pubs/INSCI2016.pdf},\n keywords = {Decision support, Indicators, Heterogeneous web resources, SME internationalisation, Semantic integration, SPARQL, statistics ontologies, CUBE},\n doi = {10.1007/978-3-319-45982-0_18},\n abstract = {Given the current economic situation and the financial crisis in many European countries, Small and Medium Enterprises (SMEs) have found interna- tionalisation and exportation of their products as the main way out of this crisis. In this paper, we provide a decision support system that semantically aggregates information from many heterogeneous web resources and provides guidance to SMEs for their potential investments. The main contributions of this paper are the introduction of SME internationalisation indicators that can be considered for such decisions, as well as the novel decision support system for SME inter- nationalisation based on inference over semantically integrated data from heterogeneous web resources. The system is evaluated by SME experts in realistic scenarios in the section of dairy products.},\n session = {13 Sep 14:20: Smart Cities and Data Analysis Issues},\n}\n\n
\n
\n\n\n
\n Given the current economic situation and the financial crisis in many European countries, Small and Medium Enterprises (SMEs) have found interna- tionalisation and exportation of their products as the main way out of this crisis. In this paper, we provide a decision support system that semantically aggregates information from many heterogeneous web resources and provides guidance to SMEs for their potential investments. The main contributions of this paper are the introduction of SME internationalisation indicators that can be considered for such decisions, as well as the novel decision support system for SME inter- nationalisation based on inference over semantically integrated data from heterogeneous web resources. The system is evaluated by SME experts in realistic scenarios in the section of dairy products.\n
\n\n\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Domain-specific modeling: Towards a Food and Drink Gazetteer.\n \n \n \n \n\n\n \n Tagarev, A.; Tolosi, L.; and Alexiev, V.\n\n\n \n\n\n\n In Cardoso, J.; Guerra, F.; Houben, G.; Pinto, A. M.; and Velegrakis, Y., editor(s),
Semantic Keyword-based Search on Structured Data Sources, volume 9398, of
Lecture Notes in Computer Science, pages 182-196, January 2016. Springer\n
First COST Action IC1302 International KEYSTONE Conference (IKC 2015), Coimbra, Portugal, September 8-9, 2015. Revised Selected Papers\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n \n slides\n \n \n \n preprint\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@InProceedings{TagarevTolosiAlexiev2017-FD,\n author = {Andrey Tagarev and Laura Tolosi and Vladimir Alexiev},\n title = {{Domain-specific modeling: Towards a Food and Drink Gazetteer}},\n booktitle = {{Semantic Keyword-based Search on Structured Data Sources}},\n year = 2016,\n editor = {Jorge Cardoso and Francesco Guerra and Geert-Jan Houben and Alexandre Miguel Pinto and Yannis Velegrakis},\n volume = 9398,\n series = {Lecture Notes in Computer Science},\n pages = {182-196},\n month = jan,\n publisher = {Springer},\n note = {First COST Action IC1302 International KEYSTONE Conference (IKC 2015), Coimbra, Portugal, September 8-9, 2015. Revised Selected Papers},\n url = {https://link.springer.com/chapter/10.1007/978-3-319-27932-9_16},\n url_Slides = {https://rawgit2.com/VladimirAlexiev/my/master/pubs/Tagarev2015-DomainSpecificGazetteer-slides.pdf},\n url_Preprint = {https://rawgit2.com/VladimirAlexiev/my/master/pubs/Tagarev2015-DomainSpecificGazetteer.pdf},\n keywords = {classification, categorization, Wikipedia, DBpedia, gazetteer, Europeana, Cultural Heritage, concept extraction, semantic enrichment, food and drink},\n chapter = 16,\n doi = {10.1007/978-3-319-27932-9_16},\n isbn = {978-3-319-27932-9},\n abstract = {Our goal is to build a Food and Drink (FD) gazetteer that can serve for classification of general, FD-related concepts, efficient faceted search or automated semantic enrichment. Fully supervised design of a domain-specific models "ex novo" is not scalable. Integration of several ready knowledge bases is tedious and does not ensure coverage. Completely data-driven approaches require a large amount of training data, which is not always available. In cases when the domain is not very specific (as the FD domain), re-using encyclopedic knowledge bases like Wikipedia may be a good idea. We propose here a semi-supervised approach, that uses a restricted Wikipedia as a base for the modeling, achieved by selecting a domain-relevant Wikipedia category as root for the model and all its subcategories, combined with expert and data-driven pruning of irrelevant categories.},\n}\n\n
\n
\n\n\n
\n Our goal is to build a Food and Drink (FD) gazetteer that can serve for classification of general, FD-related concepts, efficient faceted search or automated semantic enrichment. Fully supervised design of a domain-specific models \"ex novo\" is not scalable. Integration of several ready knowledge bases is tedious and does not ensure coverage. Completely data-driven approaches require a large amount of training data, which is not always available. In cases when the domain is not very specific (as the FD domain), re-using encyclopedic knowledge bases like Wikipedia may be a good idea. We propose here a semi-supervised approach, that uses a restricted Wikipedia as a base for the modeling, achieved by selecting a domain-relevant Wikipedia category as root for the model and all its subcategories, combined with expert and data-driven pruning of irrelevant categories.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Linked Open Data for Cultural Heritage Institutions: Build Narratives through Connecting Artifacts.\n \n \n \n \n\n\n \n Uzunov, I.; and Alexiev, V.\n\n\n \n\n\n\n In
Museum Exhibits and Standards: A Look Ahead, Sofia, Bulgaria, November 2016. Bulgarian-American Fulbright Commission for Educational Exchange: Bi-National Commission for the Preservation of Bulgaria's Cultural Heritage\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n \n pdf\n \n \n \n slides\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@InProceedings{UzunovAlexiev2016-Fulbright,\n author = {Ilian Uzunov and Vladimir Alexiev},\n title = {{Linked Open Data for Cultural Heritage Institutions: Build Narratives through Connecting Artifacts}},\n booktitle = {{Museum Exhibits and Standards: A Look Ahead}},\n year = 2016,\n month = nov,\n address = {Sofia, Bulgaria},\n organization = {Bulgarian-American Fulbright Commission for Educational Exchange: Bi-National Commission for the Preservation of Bulgaria's Cultural Heritage},\n url = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20161128-fulbright/index-full.html},\n url_PDF = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20161128-fulbright/Linked_Open_Data_for_Cultural_Heritage_Institutions.pdf},\n url_Slides = {https://rawgit2.com/VladimirAlexiev/my/master/pres/20161128-fulbright/index.html},\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n