\n \n \n
\n
\n\n \n \n \n \n \n Towards Extreme and Sustainable Graph Processing for Urgent Societal Challenges in Europe.\n \n \n \n\n\n \n Prodan, R.; Kimovski, D.; Bartolini, A.; Cochez, M.; Iosup, A.; Kharlamov, E.; Rožanec, J.; Vasiliu, L.; and Vărbănescu, A. L.\n\n\n \n\n\n\n In
2022 IEEE Cloud Summit, pages 23–30, October 2022. \n
\n\n
\n\n
\n\n
\n\n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{prodan_towards_2022,\n\ttitle = {Towards {Extreme} and {Sustainable} {Graph} {Processing} for {Urgent} {Societal} {Challenges} in {Europe}},\n\tdoi = {10.1109/CloudSummit54781.2022.00010},\n\tabstract = {The Graph-Massivizer project, funded by the Horizon Europe research and innovation program, researches and develops a high-performance, scalable, and sustainable platform for information processing and reasoning based on the massive graph (MG) representation of extreme data. It delivers a toolkit of five open-source software tools and FAIR graph datasets covering the sustainable lifecycle of processing extreme data as MGs. The tools focus on holistic usability (from extreme data ingestion and MG creation), automated intelligence (through analytics and reasoning), performance modelling, and environmental sustainability tradeoffs, supported by credible data-driven evidence across the computing continuum. The automated operation uses the emerging serverless computing paradigm for efficiency and event responsiveness. Thus, it supports experienced and novice stakeholders from a broad group of large and small organisations to capitalise on extreme data through MG programming and processing. Graph-Massivizer validates its innovation on four complementary use cases considering their extreme data properties and coverage of the three sustainability pillars (economy, society, and environment): sustainable green finance, global environment protection foresight, green AI for the sustainable automotive industry, and data centre digital twin for exascale computing. Graph-Massivizer promises 70\\% more efficient analytics than AliGraph, and 30 \\% improved energy awareness for extract, transform and load storage operations than Amazon Redshift. Furthermore, it aims to demonstrate a possible two-fold improvement in data centre energy efficiency and over 25 \\% lower greenhouse gas emissions for basic graph operations.},\n\tbooktitle = {2022 {IEEE} {Cloud} {Summit}},\n\tauthor = {Prodan, Radu and Kimovski, Dragi and Bartolini, Andrea and Cochez, Michael and Iosup, Alexandru and Kharlamov, Evgeny and Rožanec, Jože and Vasiliu, Laurenţiu and Vărbănescu, Ana Lucia},\n\tmonth = oct,\n\tyear = {2022},\n\tkeywords = {Cognition, Data centers, Europe, Extreme data, Green products, Serverless computing, Technological innovation, Transforms, graph processing, serverless computing, sustainability},\n\tpages = {23--30},\n}\n\n
\n
\n\n\n
\n The Graph-Massivizer project, funded by the Horizon Europe research and innovation program, researches and develops a high-performance, scalable, and sustainable platform for information processing and reasoning based on the massive graph (MG) representation of extreme data. It delivers a toolkit of five open-source software tools and FAIR graph datasets covering the sustainable lifecycle of processing extreme data as MGs. The tools focus on holistic usability (from extreme data ingestion and MG creation), automated intelligence (through analytics and reasoning), performance modelling, and environmental sustainability tradeoffs, supported by credible data-driven evidence across the computing continuum. The automated operation uses the emerging serverless computing paradigm for efficiency and event responsiveness. Thus, it supports experienced and novice stakeholders from a broad group of large and small organisations to capitalise on extreme data through MG programming and processing. Graph-Massivizer validates its innovation on four complementary use cases considering their extreme data properties and coverage of the three sustainability pillars (economy, society, and environment): sustainable green finance, global environment protection foresight, green AI for the sustainable automotive industry, and data centre digital twin for exascale computing. Graph-Massivizer promises 70% more efficient analytics than AliGraph, and 30 % improved energy awareness for extract, transform and load storage operations than Amazon Redshift. Furthermore, it aims to demonstrate a possible two-fold improvement in data centre energy efficiency and over 25 % lower greenhouse gas emissions for basic graph operations.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Intersection of Parallels as an Early Stopping Criterion.\n \n \n \n \n\n\n \n Vardasbi, A.; de Rijke, M.; and Dehghani, M.\n\n\n \n\n\n\n In
Proceedings of the 31st ACM International Conference on Information & Knowledge Management, of
CIKM '22, pages 1965–1974, New York, NY, USA, October 2022. Association for Computing Machinery\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{vardasbi_intersection_2022,\n\taddress = {New York, NY, USA},\n\tseries = {{CIKM} '22},\n\ttitle = {Intersection of {Parallels} as an {Early} {Stopping} {Criterion}},\n\tisbn = {978-1-4503-9236-5},\n\turl = {https://dl.acm.org/doi/10.1145/3511808.3557366},\n\tdoi = {10.1145/3511808.3557366},\n\tabstract = {A common way to avoid overfitting in supervised learning is early stopping, where a held-out set is used for iterative evaluation during training to find a sweet spot in the number of training steps that gives maximum generalization. However, such a method requires a disjoint validation set, thus part of the labeled data from the training set is usually left out for this purpose, which is not ideal when training data is scarce. Furthermore, when the training labels are noisy, the performance of the model over a validation set may not be an accurate proxy for generalization. In this paper, we propose a method to spot an early stopping point in the training iterations of an overparameterized (NN) without the need for a validation set. We first show that in the overparameterized regime the randomly initialized weights of a linear model converge to the same direction during training. Using this result, we propose to train two parallel instances of a linear model, initialized with different random seeds, and use their intersection as a signal to detect overfitting. In order to detect intersection, we use the cosine distance between the weights of the parallel models during training iterations. Noticing that the final layer of a NN is a linear map of pre-last layer activations to output logits, we build on our criterion for linear models and propose an extension to multi-layer networks, using the new notion of counterfactual weights. We conduct experiments on two areas that early stopping has noticeable impact on preventing overfitting of a NN: (i) learning from noisy labels; and (ii) learning to rank in information retrieval. Our experiments on four widely used datasets confirm the effectiveness of our method for generalization. For a wide range of learning rates, our method, called Cosine-Distance Criterion (CDC), leads to better generalization on average than all the methods that we compare against in almost all of the tested cases.},\n\turldate = {2023-04-21},\n\tbooktitle = {Proceedings of the 31st {ACM} {International} {Conference} on {Information} \\& {Knowledge} {Management}},\n\tpublisher = {Association for Computing Machinery},\n\tauthor = {Vardasbi, Ali and de Rijke, Maarten and Dehghani, Mostafa},\n\tmonth = oct,\n\tyear = {2022},\n\tkeywords = {cosine distance, early stopping, generalization, overparameterization},\n\tpages = {1965--1974},\n}\n\n
\n
\n\n\n
\n A common way to avoid overfitting in supervised learning is early stopping, where a held-out set is used for iterative evaluation during training to find a sweet spot in the number of training steps that gives maximum generalization. However, such a method requires a disjoint validation set, thus part of the labeled data from the training set is usually left out for this purpose, which is not ideal when training data is scarce. Furthermore, when the training labels are noisy, the performance of the model over a validation set may not be an accurate proxy for generalization. In this paper, we propose a method to spot an early stopping point in the training iterations of an overparameterized (NN) without the need for a validation set. We first show that in the overparameterized regime the randomly initialized weights of a linear model converge to the same direction during training. Using this result, we propose to train two parallel instances of a linear model, initialized with different random seeds, and use their intersection as a signal to detect overfitting. In order to detect intersection, we use the cosine distance between the weights of the parallel models during training iterations. Noticing that the final layer of a NN is a linear map of pre-last layer activations to output logits, we build on our criterion for linear models and propose an extension to multi-layer networks, using the new notion of counterfactual weights. We conduct experiments on two areas that early stopping has noticeable impact on preventing overfitting of a NN: (i) learning from noisy labels; and (ii) learning to rank in information retrieval. Our experiments on four widely used datasets confirm the effectiveness of our method for generalization. For a wide range of learning rates, our method, called Cosine-Distance Criterion (CDC), leads to better generalization on average than all the methods that we compare against in almost all of the tested cases.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Probabilistic Permutation Graph Search: Black-Box Optimization for Fairness in Ranking.\n \n \n \n \n\n\n \n Vardasbi, A.; Sarvi, F.; and de Rijke, M.\n\n\n \n\n\n\n In
Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, of
SIGIR '22, pages 715–725, New York, NY, USA, July 2022. Association for Computing Machinery\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{vardasbi_probabilistic_2022,\n\taddress = {New York, NY, USA},\n\tseries = {{SIGIR} '22},\n\ttitle = {Probabilistic {Permutation} {Graph} {Search}: {Black}-{Box} {Optimization} for {Fairness} in {Ranking}},\n\tisbn = {978-1-4503-8732-3},\n\tshorttitle = {Probabilistic {Permutation} {Graph} {Search}},\n\turl = {https://doi.org/10.1145/3477495.3532045},\n\tdoi = {10.1145/3477495.3532045},\n\tabstract = {There are several measures for fairness in ranking, based on different underlying assumptions and perspectives. {\\textbackslash}acPL optimization with the REINFORCE algorithm can be used for optimizing black-box objective functions over permutations. In particular, it can be used for optimizing fairness measures. However, though effective for queries with a moderate number of repeating sessions, {\\textbackslash}acPL optimization has room for improvement for queries with a small number of repeating sessions. In this paper, we present a novel way of representing permutation distributions, based on the notion of permutation graphs. Similar to{\\textasciitilde}{\\textbackslash}acPL, our distribution representation, called{\\textasciitilde}{\\textbackslash}acPPG, can be used for black-box optimization of fairness. Different from{\\textasciitilde}{\\textbackslash}acPL, where pointwise logits are used as the distribution parameters, in{\\textasciitilde}{\\textbackslash}acPPG pairwise inversion probabilities together with a reference permutation construct the distribution. As such, the reference permutation can be set to the best sampled permutation regarding the objective function, making{\\textasciitilde}{\\textbackslash}acPPG suitable for both deterministic and stochastic rankings. Our experiments show that{\\textasciitilde}{\\textbackslash}acPPG, while comparable to{\\textasciitilde}{\\textbackslash}acPL for larger session repetitions (i.e., stochastic ranking), improves over{\\textasciitilde}{\\textbackslash}acPL for optimizing fairness metrics for queries with one session (i.e., deterministic ranking). Additionally, when accurate utility estimations are available, e.g., in tabular models, the performance of {\\textbackslash}acPPG in fairness optimization is significantly boosted compared to lower quality utility estimations from a learning to rank model, leading to a large performance gap with PL. Finally, the pairwise probabilities make it possible to impose pairwise constraints such as "item \\$d\\_1\\$ should always be ranked higher than item \\$d\\_2\\$.'' Such constraints can be used to simultaneously optimize the fairness metric and control another objective such as ranking performance.},\n\turldate = {2023-04-21},\n\tbooktitle = {Proceedings of the 45th {International} {ACM} {SIGIR} {Conference} on {Research} and {Development} in {Information} {Retrieval}},\n\tpublisher = {Association for Computing Machinery},\n\tauthor = {Vardasbi, Ali and Sarvi, Fatemeh and de Rijke, Maarten},\n\tmonth = jul,\n\tyear = {2022},\n\tkeywords = {fairness in ranking, permutation distribution, permutation graph, plackett-luce},\n\tpages = {715--725},\n}\n\n
\n
\n\n\n
\n There are several measures for fairness in ranking, based on different underlying assumptions and perspectives. \\acPL optimization with the REINFORCE algorithm can be used for optimizing black-box objective functions over permutations. In particular, it can be used for optimizing fairness measures. However, though effective for queries with a moderate number of repeating sessions, \\acPL optimization has room for improvement for queries with a small number of repeating sessions. In this paper, we present a novel way of representing permutation distributions, based on the notion of permutation graphs. Similar to~\\acPL, our distribution representation, called~\\acPPG, can be used for black-box optimization of fairness. Different from~\\acPL, where pointwise logits are used as the distribution parameters, in~\\acPPG pairwise inversion probabilities together with a reference permutation construct the distribution. As such, the reference permutation can be set to the best sampled permutation regarding the objective function, making~\\acPPG suitable for both deterministic and stochastic rankings. Our experiments show that~\\acPPG, while comparable to~\\acPL for larger session repetitions (i.e., stochastic ranking), improves over~\\acPL for optimizing fairness metrics for queries with one session (i.e., deterministic ranking). Additionally, when accurate utility estimations are available, e.g., in tabular models, the performance of \\acPPG in fairness optimization is significantly boosted compared to lower quality utility estimations from a learning to rank model, leading to a large performance gap with PL. Finally, the pairwise probabilities make it possible to impose pairwise constraints such as \"item $d_1$ should always be ranked higher than item $d_2$.'' Such constraints can be used to simultaneously optimize the fairness metric and control another objective such as ranking performance.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Recommending scientific datasets using author networks in ensemble methods.\n \n \n \n \n\n\n \n Wang, X.; van Harmelen, F.; and Huang, Z.\n\n\n \n\n\n\n
Data Science, 5(2): 167–193. July 2022.\n
Publisher: IOS Press\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{wang_recommending_2022,\n\ttitle = {Recommending scientific datasets using author networks in ensemble methods},\n\tvolume = {5},\n\tissn = {2451-8484},\n\turl = {https://content.iospress.com/articles/data-science/ds220056},\n\tdoi = {10.3233/DS-220056},\n\tabstract = {Open access to datasets is increasingly driving modern science. Consequently, discovering such datasets is becoming an important functionality for scientists in many different fields. We investigate methods for dataset recommendation : the task of re},\n\tlanguage = {en},\n\tnumber = {2},\n\turldate = {2023-04-22},\n\tjournal = {Data Science},\n\tauthor = {Wang, Xu and van Harmelen, Frank and Huang, Zhisheng},\n\tmonth = jul,\n\tyear = {2022},\n\tnote = {Publisher: IOS Press},\n\tpages = {167--193},\n}\n\n
\n
\n\n\n
\n Open access to datasets is increasingly driving modern science. Consequently, discovering such datasets is becoming an important functionality for scientists in many different fields. We investigate methods for dataset recommendation : the task of re\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n Hyperbolic Embedding Inference for Structured Multi-Label Prediction.\n \n \n \n\n\n \n Xiong, B.; Cochez, M.; Nayyeri, M.; and Staab, S.\n\n\n \n\n\n\n In
NeurIPS2022, November 2022. \n
\n\n
\n\n
\n\n
\n\n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{xiong_hyperbolic_2022,\n\ttitle = {Hyperbolic {Embedding} {Inference} for {Structured} {Multi}-{Label} {Prediction}},\n\tbooktitle = {{NeurIPS2022}},\n\tauthor = {Xiong, Bo and Cochez, Michael and Nayyeri, Mojtaba and Staab, Steffen},\n\tmonth = nov,\n\tyear = {2022},\n}\n\n
\n
\n\n\n\n
\n\n\n \n\n\n
\n
\n\n \n \n \n \n \n Scientific Item Recommendation Using a Citation Network.\n \n \n \n\n\n \n Wang, X.; van Harmelen, F.; Cochez, M.; Huang, Z.; Yang, B.; Kong, L.; Zhang, T.; and Qiu, M.\n\n\n \n\n\n\n In
Knowledge Science, Engineering and Management, pages 469–484, Cham, August 2022. Springer International Publishing\n
\n\n
\n\n
\n\n
\n\n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{wang_scientific_2022,\n\taddress = {Cham},\n\ttitle = {Scientific {Item} {Recommendation} {Using} a {Citation} {Network}},\n\tisbn = {978-3-031-10986-7},\n\tbooktitle = {Knowledge {Science}, {Engineering} and {Management}},\n\tpublisher = {Springer International Publishing},\n\tauthor = {Wang, Xu and van Harmelen, Frank and Cochez, Michael and Huang, Zhisheng", editor="Memmi, Gerard and Yang, Baijian and Kong, Linghe and Zhang, Tianwei and Qiu, Meikang},\n\tmonth = aug,\n\tyear = {2022},\n\tpages = {469--484},\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n Towards a standard-based open data ecosystem: analysis of DCAT-AP use at national and European level.\n \n \n \n\n\n \n Barthelemy, F.; Cochez, M.; Dimitriadis, I.; Karim, N.; Loutas, N.; Magnisalis, I.; Comet, L. M.; Peristeras, V.; and Wyns, B.\n\n\n \n\n\n\n
Electronic Government, an International Journal, 18(2): 137–180. January 2022.\n
Publisher: Inderscience Publishers (IEL)\n\n
\n\n
\n\n
\n\n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{barthelemy_towards_2022,\n\ttitle = {Towards a standard-based open data ecosystem: analysis of {DCAT}-{AP} use at national and {European} level},\n\tvolume = {18},\n\tnumber = {2},\n\tjournal = {Electronic Government, an International Journal},\n\tauthor = {Barthelemy, Florian and Cochez, Michael and Dimitriadis, Iraklis and Karim, Naila and Loutas, Nikolaos and Magnisalis, Ioannis and Comet, Lina Molinas and Peristeras, Vassilios and Wyns, Brecht},\n\tmonth = jan,\n\tyear = {2022},\n\tnote = {Publisher: Inderscience Publishers (IEL)},\n\tpages = {137--180},\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Scaling R-GCN Training with Graph Summarization.\n \n \n \n \n\n\n \n Generale, A.; Blume, T.; and Cochez, M.\n\n\n \n\n\n\n In
WWW '22: Companion Proceedings of the Web Conference 2022, of
WWW '22, pages 1073–1082, Virtual Event, Lyon, France, April 2022. Association for Computing Machinery\n
arXiv preprint arXiv:2203.02622\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{generale_scaling_2022,\n\taddress = {Virtual Event, Lyon, France},\n\tseries = {{WWW} '22},\n\ttitle = {Scaling {R}-{GCN} {Training} with {Graph} {Summarization}},\n\tisbn = {978-1-4503-9130-6},\n\turl = {https://arxiv.org/abs/2203.02622},\n\tdoi = {https://doi.org/10.1145/3487553.3524719},\n\tbooktitle = {{WWW} '22: {Companion} {Proceedings} of the {Web} {Conference} 2022},\n\tpublisher = {Association for Computing Machinery},\n\tauthor = {Generale, Alessandro and Blume, Till and Cochez, Michael},\n\tmonth = apr,\n\tyear = {2022},\n\tnote = {arXiv preprint arXiv:2203.02622},\n\tkeywords = {graph neural network, graph summarization, scalability},\n\tpages = {1073--1082},\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Exposure-Aware Recommendation using Contextual Bandits.\n \n \n \n \n\n\n \n Mansoury, M.; Mobasher, B.; and van Hoof, H.\n\n\n \n\n\n\n In September 2022. 5th FAccTRec Workshop on Responsible Recommendation in conjunction with ACM RecSys 2022\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 4 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{mansoury_exposure-aware_2022,\n\ttitle = {Exposure-{Aware} {Recommendation} using {Contextual} {Bandits}},\n\tcopyright = {Creative Commons Attribution 4.0 International},\n\turl = {https://arxiv.org/abs/2209.01665},\n\tdoi = {10.48550/ARXIV.2209.01665},\n\tpublisher = {5th FAccTRec Workshop on Responsible Recommendation in conjunction with ACM RecSys 2022},\n\tauthor = {Mansoury, Masoud and Mobasher, Bamshad and van Hoof, Herke},\n\tmonth = sep,\n\tyear = {2022},\n\tkeywords = {FOS: Computer and information sciences, Information Retrieval (cs.IR)},\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Query Embedding on Hyper-Relational Knowledge Graphs.\n \n \n \n \n\n\n \n Alivanistos, D.; Berrendorf, M.; Cochez, M.; and Galkin, M.\n\n\n \n\n\n\n In
The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022, April 2022. OpenReview.net\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{alivanistos_query_2022,\n\ttitle = {Query {Embedding} on {Hyper}-{Relational} {Knowledge} {Graphs}},\n\turl = {https://openreview.net/forum?id=4rLw09TgRw9},\n\tbooktitle = {The {Tenth} {International} {Conference} on {Learning} {Representations}, {ICLR} 2022, {Virtual} {Event}, {April} 25-29, 2022},\n\tpublisher = {OpenReview.net},\n\tauthor = {Alivanistos, Dimitrios and Berrendorf, Max and Cochez, Michael and Galkin, Mikhail},\n\tmonth = apr,\n\tyear = {2022},\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Potential Energy to Improve Link Prediction With Relational Graph Neural Networks.\n \n \n \n \n\n\n \n Colombo, S.; Alivanistos, D.; and Cochez, M.\n\n\n \n\n\n\n In Martin, A.; Hinkelmann, K.; Fill, H.; Gerber, A.; Lenat, D.; Stolle, R.; and Harmelen, F. v., editor(s),
Proceedings of the AAAI 2022 Spring Symposium on Machine Learning and Knowledge Engineering for Hybrid Intelligence (AAAI-MAKE 2022), Stanford University, Palo Alto, California, USA, March 21-23, 2022, volume 3121, of
CEUR Workshop Proceedings, March 2022. CEUR-WS.org\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{colombo_potential_2022,\n\tseries = {{CEUR} {Workshop} {Proceedings}},\n\ttitle = {Potential {Energy} to {Improve} {Link} {Prediction} {With} {Relational} {Graph} {Neural} {Networks}},\n\tvolume = {3121},\n\turl = {http://ceur-ws.org/Vol-3121/short2.pdf},\n\tbooktitle = {Proceedings of the {AAAI} 2022 {Spring} {Symposium} on {Machine} {Learning} and {Knowledge} {Engineering} for {Hybrid} {Intelligence} ({AAAI}-{MAKE} 2022), {Stanford} {University}, {Palo} {Alto}, {California}, {USA}, {March} 21-23, 2022},\n\tpublisher = {CEUR-WS.org},\n\tauthor = {Colombo, Simone and Alivanistos, Dimitrios and Cochez, Michael},\n\teditor = {Martin, Andreas and Hinkelmann, Knut and Fill, Hans-Georg and Gerber, Aurona and Lenat, Doug and Stolle, Reinhard and Harmelen, Frank van},\n\tmonth = mar,\n\tyear = {2022},\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Prompting as Probing: Using Language Models for Knowledge Base Construction.\n \n \n \n \n\n\n \n Alivanistos, D.; Santamaría, S. B.; Cochez, M.; Kalo, J.; van Krieken, E.; and Thanapalasingam, T.\n\n\n \n\n\n\n In
Proceedings of the Semantic Web Challenge on Knowledge Base Construction from Pre-trained Language Models 2022 co-located with the 21st International Semantic Web Conference (ISWC2022), volume 3274, pages 11–34, October 2022. CEUR-ws.org\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 4 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{alivanistos_prompting_2022,\n\ttitle = {Prompting as {Probing}: {Using} {Language} {Models} for {Knowledge} {Base} {Construction}},\n\tvolume = {3274},\n\tcopyright = {arXiv.org perpetual, non-exclusive license},\n\turl = {https://ceur-ws.org/Vol-3274/paper2.pdf},\n\tdoi = {10.48550/ARXIV.2208.11057},\n\tbooktitle = {Proceedings of the {Semantic} {Web} {Challenge} on {Knowledge} {Base} {Construction} from {Pre}-trained {Language} {Models} 2022 co-located with the 21st {International} {Semantic} {Web} {Conference} ({ISWC2022})},\n\tpublisher = {CEUR-ws.org},\n\tauthor = {Alivanistos, Dimitrios and Santamaría, Selene Báez and Cochez, Michael and Kalo, Jan-Christoph and van Krieken, Emile and Thanapalasingam, Thiviyan},\n\tmonth = oct,\n\tyear = {2022},\n\tkeywords = {Artificial Intelligence (cs.AI), Computation and Language (cs.CL), FOS: Computer and information sciences},\n\tpages = {11--34},\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n The Seventh Workshop on Search-Oriented Conversational Artificial Intelligence (SCAI'22).\n \n \n \n \n\n\n \n Penha, G.; Vakulenko, S.; Dusek, O.; Clark, L.; Pal, V.; and Adlakha, V.\n\n\n \n\n\n\n In
Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, of
SIGIR '22, pages 3466–3469, New York, NY, USA, July 2022. Association for Computing Machinery\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 3 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{penha_seventh_2022,\n\taddress = {New York, NY, USA},\n\tseries = {{SIGIR} '22},\n\ttitle = {The {Seventh} {Workshop} on {Search}-{Oriented} {Conversational} {Artificial} {Intelligence} ({SCAI}'22)},\n\tisbn = {978-1-4503-8732-3},\n\turl = {https://dl.acm.org/doi/10.1145/3477495.3531700},\n\tdoi = {10.1145/3477495.3531700},\n\tabstract = {The goal of the seventh edition of SCAI (https://scai.info) is to bring together and further grow a community of researchers and practitioners interested in conversational systems for information access. The previous iterations of the workshop already demonstrated the breadth and multidisciplinarity inherent in the design and development of conversational search agents. The proposed shift from traditional web search to search interfaces enabled via human-like dialogue leads to a number of challenges, and although such challenges have received more attention in the recent years, there are many pending research questions that should be addressed by the information retrieval community and can largely benefit from a collaboration with other research fields, such as natural language processing, machine learning, human-computer interaction and dialogue systems. This workshop is intended as a platform enabling a continuous discussion of the major research challenges that surround the design of search-oriented conversational systems. This year, participants have the opportunity to meet in person and have more in-depth interactive discussions with a full-day onsite workshop.},\n\turldate = {2023-04-20},\n\tbooktitle = {Proceedings of the 45th {International} {ACM} {SIGIR} {Conference} on {Research} and {Development} in {Information} {Retrieval}},\n\tpublisher = {Association for Computing Machinery},\n\tauthor = {Penha, Gustavo and Vakulenko, Svitlana and Dusek, Ondrej and Clark, Leigh and Pal, Vaishali and Adlakha, Vaibhav},\n\tmonth = jul,\n\tyear = {2022},\n\tkeywords = {conversational information access, conversational search, information-seeking dialogue},\n\tpages = {3466--3469},\n}\n\n
\n
\n\n\n
\n The goal of the seventh edition of SCAI (https://scai.info) is to bring together and further grow a community of researchers and practitioners interested in conversational systems for information access. The previous iterations of the workshop already demonstrated the breadth and multidisciplinarity inherent in the design and development of conversational search agents. The proposed shift from traditional web search to search interfaces enabled via human-like dialogue leads to a number of challenges, and although such challenges have received more attention in the recent years, there are many pending research questions that should be addressed by the information retrieval community and can largely benefit from a collaboration with other research fields, such as natural language processing, machine learning, human-computer interaction and dialogue systems. This workshop is intended as a platform enabling a continuous discussion of the major research challenges that surround the design of search-oriented conversational systems. This year, participants have the opportunity to meet in person and have more in-depth interactive discussions with a full-day onsite workshop.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n MORS 2022: The Second Workshop on Multi-Objective Recommender Systems.\n \n \n \n \n\n\n \n Abdollahpouri, H.; Sahebi, S.; Elahi, M.; Mansoury, M.; Loni, B.; Nazari, Z.; and Dimakopoulou, M.\n\n\n \n\n\n\n In
Sixteenth ACM Conference on Recommender Systems, pages 658–660, Seattle WA USA, September 2022. ACM\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 2 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{abdollahpouri_mors_2022,\n\taddress = {Seattle WA USA},\n\ttitle = {{MORS} 2022: {The} {Second} {Workshop} on {Multi}-{Objective} {Recommender} {Systems}},\n\tisbn = {978-1-4503-9278-5},\n\tshorttitle = {{MORS} 2022},\n\turl = {https://dl.acm.org/doi/10.1145/3523227.3547410},\n\tdoi = {10.1145/3523227.3547410},\n\tlanguage = {en},\n\turldate = {2023-04-20},\n\tbooktitle = {Sixteenth {ACM} {Conference} on {Recommender} {Systems}},\n\tpublisher = {ACM},\n\tauthor = {Abdollahpouri, Himan and Sahebi, Shaghayegh and Elahi, Mehdi and Mansoury, Masoud and Loni, Babak and Nazari, Zahra and Dimakopoulou, Maria},\n\tmonth = sep,\n\tyear = {2022},\n\tpages = {658--660},\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n SlotGAN: Detecting Mentions in Text via Adversarial Distant Learning.\n \n \n \n \n\n\n \n Daza, D.; Cochez, M.; and Groth, P.\n\n\n \n\n\n\n In
Proceedings of the Sixth Workshop on Structured Prediction for NLP, pages 32–39, Dublin, Ireland, May 2022. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{daza_slotgan_2022,\n\taddress = {Dublin, Ireland},\n\ttitle = {{SlotGAN}: {Detecting} {Mentions} in {Text} via {Adversarial} {Distant} {Learning}},\n\turl = {https://aclanthology.org/2022.spnlp-1.4},\n\tdoi = {10.18653/v1/2022.spnlp-1.4},\n\tabstract = {We present SlotGAN, a framework for training a mention detection model that only requires unlabeled text and a gazetteer. It consists of a generator trained to extract spans from an input sentence, and a discriminator trained to determine whether a span comes from the generator, or from the gazetteer.We evaluate the method on English newswire data and compare it against supervised, weakly-supervised, and unsupervised methods. We find that the performance of the method is lower than these baselines, because it tends to generate more and longer spans, and in some cases it relies only on capitalization. In other cases, it generates spans that are valid but differ from the benchmark. When evaluated with metrics based on overlap, we find that SlotGAN performs within 95\\% of the precision of a supervised method, and 84\\% of its recall. Our results suggest that the model can generate spans that overlap well, but an additional filtering mechanism is required.},\n\tbooktitle = {Proceedings of the {Sixth} {Workshop} on {Structured} {Prediction} for {NLP}},\n\tpublisher = {Association for Computational Linguistics},\n\tauthor = {Daza, Daniel and Cochez, Michael and Groth, Paul},\n\tmonth = may,\n\tyear = {2022},\n\tpages = {32--39},\n}\n\n
\n
\n\n\n
\n We present SlotGAN, a framework for training a mention detection model that only requires unlabeled text and a gazetteer. It consists of a generator trained to extract spans from an input sentence, and a discriminator trained to determine whether a span comes from the generator, or from the gazetteer.We evaluate the method on English newswire data and compare it against supervised, weakly-supervised, and unsupervised methods. We find that the performance of the method is lower than these baselines, because it tends to generate more and longer spans, and in some cases it relies only on capitalization. In other cases, it generates spans that are valid but differ from the benchmark. When evaluated with metrics based on overlap, we find that SlotGAN performs within 95% of the precision of a supervised method, and 84% of its recall. Our results suggest that the model can generate spans that overlap well, but an additional filtering mechanism is required.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Visualising the effects of ontology changes and studying their understanding with ChImp.\n \n \n \n \n\n\n \n Pernisch, R.; Dell’Aglio, D.; Serbak, M.; Gonçalves, R. S.; and Bernstein, A.\n\n\n \n\n\n\n
Journal of Web Semantics, 74: 100715. October 2022.\n
\n\n
\n\n
\n\n
\n\n \n \n Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 7 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@article{pernisch_visualising_2022,\n\ttitle = {Visualising the effects of ontology changes and studying their understanding with {ChImp}},\n\tvolume = {74},\n\tissn = {1570-8268},\n\turl = {https://www.sciencedirect.com/science/article/pii/S1570826822000117},\n\tdoi = {10.1016/j.websem.2022.100715},\n\tabstract = {Due to the Semantic Web’s decentralised nature, ontology engineers rarely know all applications that leverage their ontology. Consequently, they are unaware of the full extent of possible consequences that changes might cause to the ontology. Our goal is to lessen the gap between ontology engineers and users by investigating ontology engineers’ understanding of ontology changes’ impact at editing time. Hence, this paper introduces the Protégé plugin ChImp which we use to reach our goal. We elicited requirements for ChImp through a questionnaire with ontology engineers. We then developed ChImp according to these requirements and it displays all changes of a given session and provides selected information on said changes and their effects. For each change, it computes a number of metrics on both the ontology and its materialisation. It displays those metrics on both the originally loaded ontology at the beginning of the editing session and the current state to help ontology engineers understand the impact of their changes. We investigated the informativeness of materialisation impact measures, the meaning of severe impact, and also the usefulness of ChImp in an online user study with 36 ontology engineers. We asked the participants to solve two ontology engineering tasks – with and without ChImp (assigned in random order) – and answer in-depth questions about the applied changes as well as the materialisation impact measures. We found that ChImp increased the participants’ understanding of change effects and that they felt better informed. Answers also suggest that the proposed measures were useful and informative. We also learned that the participants consider different outcomes of changes severe, but most would define severity based on the amount of changes to the materialisation compared to its size. The participants also acknowledged the importance of quantifying the impact of changes and that the study will affect their approach of editing ontologies.},\n\tlanguage = {en},\n\turldate = {2022-06-20},\n\tjournal = {Journal of Web Semantics},\n\tauthor = {Pernisch, Romana and Dell’Aglio, Daniele and Serbak, Mirko and Gonçalves, Rafael S. and Bernstein, Abraham},\n\tmonth = oct,\n\tyear = {2022},\n\tkeywords = {Materialisation, Ontology editing, Ontology evolution impact, User study},\n\tpages = {100715},\n}\n\n
\n
\n\n\n
\n Due to the Semantic Web’s decentralised nature, ontology engineers rarely know all applications that leverage their ontology. Consequently, they are unaware of the full extent of possible consequences that changes might cause to the ontology. Our goal is to lessen the gap between ontology engineers and users by investigating ontology engineers’ understanding of ontology changes’ impact at editing time. Hence, this paper introduces the Protégé plugin ChImp which we use to reach our goal. We elicited requirements for ChImp through a questionnaire with ontology engineers. We then developed ChImp according to these requirements and it displays all changes of a given session and provides selected information on said changes and their effects. For each change, it computes a number of metrics on both the ontology and its materialisation. It displays those metrics on both the originally loaded ontology at the beginning of the editing session and the current state to help ontology engineers understand the impact of their changes. We investigated the informativeness of materialisation impact measures, the meaning of severe impact, and also the usefulness of ChImp in an online user study with 36 ontology engineers. We asked the participants to solve two ontology engineering tasks – with and without ChImp (assigned in random order) – and answer in-depth questions about the applied changes as well as the materialisation impact measures. We found that ChImp increased the participants’ understanding of change effects and that they felt better informed. Answers also suggest that the proposed measures were useful and informative. We also learned that the participants consider different outcomes of changes severe, but most would define severity based on the amount of changes to the materialisation compared to its size. The participants also acknowledged the importance of quantifying the impact of changes and that the study will affect their approach of editing ontologies.\n
\n\n\n
\n\n\n\n\n\n