var bibbase_data = {"data":"\"Loading..\"\n\n
\n\n \n\n \n\n \n \n\n \n\n \n \n\n \n\n \n
\n generated by\n \n \"bibbase.org\"\n\n \n
\n \n\n
\n\n \n\n\n
\n\n Excellent! Next you can\n create a new website with this list, or\n embed it in an existing web page by copying & pasting\n any of the following snippets.\n\n
\n JavaScript\n (easiest)\n
\n \n <script src=\"https://bibbase.org/show?bib=https://github.com/CVMLmu/cvmlmu.github.io/raw/main/cvmlmu.bib&jsonp=1&showSearch=1&jsonp=1\"></script>\n \n
\n\n PHP\n
\n \n <?php\n $contents = file_get_contents(\"https://bibbase.org/show?bib=https://github.com/CVMLmu/cvmlmu.github.io/raw/main/cvmlmu.bib&jsonp=1&showSearch=1\");\n print_r($contents);\n ?>\n \n
\n\n iFrame\n (not recommended)\n
\n \n <iframe src=\"https://bibbase.org/show?bib=https://github.com/CVMLmu/cvmlmu.github.io/raw/main/cvmlmu.bib&jsonp=1&showSearch=1\"></iframe>\n \n
\n\n

\n For more details see the documention.\n

\n
\n
\n\n
\n\n This is a preview! To use this list on your own web site\n or create a new web site from it,\n create a free account. The file will be added\n and you will be able to edit it in the File Manager.\n We will show you instructions once you've created your account.\n
\n\n
\n\n

To the site owner:

\n\n

Action required! Mendeley is changing its\n API. In order to keep using Mendeley with BibBase past April\n 14th, you need to:\n

    \n
  1. renew the authorization for BibBase on Mendeley, and
  2. \n
  3. update the BibBase URL\n in your page the same way you did when you initially set up\n this page.\n
  4. \n
\n

\n\n

\n \n \n Fix it now\n

\n
\n\n
\n\n\n
\n \n \n
\n
\n  \n 2023\n \n \n (5)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Improving GMM registration with class encoding.\n \n \n \n \n\n\n \n Panahi, S.; Chopin, J.; Ulicny, M.; and Dahyot, R.\n\n\n \n\n\n\n In Irish Machine Vision and Image Processing (IMVIP 2023), 2023. \n \n\n\n\n
\n\n\n\n \n \n \"ImprovingPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 3 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{Panahi2023,\nauthor= {Solmaz Panahi and Jeremy Chopin and Matej Ulicny and Rozenn Dahyot}, \ntitle= {Improving  GMM  registration with class encoding},\nbooktitle= {Irish Machine Vision and Image Processing (IMVIP 2023)},\nvolume= {},\nyear= {2023},\nabstract={Point set registration is critical in many applications such as  computer vision, pattern recognition, or in fields like robotics and medical imaging.\nThis paper focuses on reformulating point set registration using Gaussian Mixture Models while considering attributes associated with each point. Our approach introduces class score vectors as additional features \nto the spatial data information. By incorporating these attributes, we enhance the optimization process by penalizing incorrect matching terms. Experimental results show that our approach \nwith class scores outperforms the original algorithm  in both accuracy and speed.},\nurl= {https://github.com/solmak97/GMMReg_Extension},\ndoi={10.5281/zenodo.8205096},\nnote={},\n\n}\n\n
\n
\n\n\n
\n Point set registration is critical in many applications such as computer vision, pattern recognition, or in fields like robotics and medical imaging. This paper focuses on reformulating point set registration using Gaussian Mixture Models while considering attributes associated with each point. Our approach introduces class score vectors as additional features to the spatial data information. By incorporating these attributes, we enhance the optimization process by penalizing incorrect matching terms. Experimental results show that our approach with class scores outperforms the original algorithm in both accuracy and speed.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Query Based Acoustic Summarization for Podcasts.\n \n \n \n \n\n\n \n Kotey, S.; Dahyot, R.; and Harte, N.\n\n\n \n\n\n\n In Proc. INTERSPEECH 2023, pages 1483–1487, Dublin, Ireland, August 2023. \n \n\n\n\n
\n\n\n\n \n \n \"QueryPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n  \n \n 3 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@INPROCEEDINGS{KoteyInterSpeech2023,\n  author={Samantha Kotey and Rozenn Dahyot and Naomi Harte},\n   booktitle={Proc. INTERSPEECH 2023},\n  title={Query Based Acoustic Summarization for Podcasts}, \n  year={2023},\n  volume={},\n  number={},\n pages={1483--1487},\n  abstract={},\n  keywords={},\n   doi={10.21437/Interspeech.2023-864},\n  url={https://www.isca-speech.org/archive/pdfs/interspeech_2023/kotey23_interspeech.pdf},\n  ISSN={},\n  address={Dublin, Ireland},\n  month={August},}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Combining geolocation and height estimation of objects from street level imagery.\n \n \n \n \n\n\n \n Ulicny, M.; Krylov, V. A.; Connelly, J.; and Dahyot, R.\n\n\n \n\n\n\n Technical Report 2023.\n \n\n\n\n
\n\n\n\n \n \n \"CombiningPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 2 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@techreport{ulicny2023combining,\n       doi={10.48550/arXiv.2305.08232},\n\t   url={https://arxiv.org/pdf/2305.08232.pdf},\n\t   abstract={We propose a pipeline for combined multi-class object geolocation and height estimation from street level RGB imagery, which is considered as a single available input data modality. Our solution is formulated via Markov Random Field optimization with deterministic output. The proposed technique uses image metadata along with coordinates of objects detected in the image plane as found by a custom-trained Convolutional Neural Network. Computing the object height using our methodology, in addition to object geolocation, has negligible effect on the overall computational cost. Accuracy is demonstrated experimentally for water drains and road signs on which we achieve average elevation estimation error lower than 20cm.},\n      title={Combining geolocation and height estimation of objects from street level imagery}, \n      author={Matej Ulicny and Vladimir A. Krylov and Julie Connelly and Rozenn Dahyot},\n      year={2023},\n      eprint={2305.08232},\n      archivePrefix={arXiv},\n      primaryClass={cs.CV}\n}\n\n
\n
\n\n\n
\n We propose a pipeline for combined multi-class object geolocation and height estimation from street level RGB imagery, which is considered as a single available input data modality. Our solution is formulated via Markov Random Field optimization with deterministic output. The proposed technique uses image metadata along with coordinates of objects detected in the image plane as found by a custom-trained Convolutional Neural Network. Computing the object height using our methodology, in addition to object geolocation, has negligible effect on the overall computational cost. Accuracy is demonstrated experimentally for water drains and road signs on which we achieve average elevation estimation error lower than 20cm.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Model-based inexact graph matching on top of DNNs for semantic scene understanding.\n \n \n \n \n\n\n \n Chopin, J.; Fasquel, J.; Mouchère, H.; Dahyot, R.; and Bloch, I.\n\n\n \n\n\n\n Computer Vision and Image Understanding,103744. 2023.\n \n\n\n\n
\n\n\n\n \n \n \"Model-basedPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@article{CHOPIN2023103744,\ntitle = {Model-based inexact graph matching on top of DNNs for semantic scene understanding},\njournal = {Computer Vision and Image Understanding},\npages = {103744},\nyear = {2023},\nissn = {1077-3142},\ndoi = {https://doi.org/10.1016/j.cviu.2023.103744},\nurl = {https://arxiv.org/pdf/2301.07468.pdf},\nauthor = {Jeremy Chopin and Jean-Baptiste Fasquel and Harold Mouchère and Rozenn Dahyot and Isabelle Bloch},\nkeywords = {Graph matching, Deep learning, Image segmentation, Volume segmentation, Quadratic assignment problem},\nabstract = {Deep learning based pipelines for semantic segmentation often ignore structural information available on annotated images used for training. We propose a novel post-processing module enforcing structural knowledge about the objects of interest to improve segmentation results provided by deep neural networks (DNNs). This module corresponds to a “many-to-one-or-none” inexact graph matching approach, and is formulated as a quadratic assignment problem. Our approach is compared to a DNN-based segmentation on two public datasets, one for face segmentation from 2D RGB images (FASSEG), and the other for brain segmentation from 3D MRIs (IBSR). Evaluations are performed using two types of structural information: distances and directional relations that are user defined, this choice being a hyper-parameter of our proposed generic framework. On FASSEG data, results show that our module improves accuracy of the DNN by about 6.3% i.e. the Hausdorff distance (HD) decreases from 22.11 to 20.71 on average. With IBSR data, the improvement is of 51% better accuracy with HD decreasing from 11.01 to 5.4. Finally, our approach is shown to be resilient to small training datasets that often limit the performance of deep learning methods: the improvement increases as the size of the training dataset decreases.}\n}\n\n
\n
\n\n\n
\n Deep learning based pipelines for semantic segmentation often ignore structural information available on annotated images used for training. We propose a novel post-processing module enforcing structural knowledge about the objects of interest to improve segmentation results provided by deep neural networks (DNNs). This module corresponds to a “many-to-one-or-none” inexact graph matching approach, and is formulated as a quadratic assignment problem. Our approach is compared to a DNN-based segmentation on two public datasets, one for face segmentation from 2D RGB images (FASSEG), and the other for brain segmentation from 3D MRIs (IBSR). Evaluations are performed using two types of structural information: distances and directional relations that are user defined, this choice being a hyper-parameter of our proposed generic framework. On FASSEG data, results show that our module improves accuracy of the DNN by about 6.3% i.e. the Hausdorff distance (HD) decreases from 22.11 to 20.71 on average. With IBSR data, the improvement is of 51% better accuracy with HD decreasing from 11.01 to 5.4. Finally, our approach is shown to be resilient to small training datasets that often limit the performance of deep learning methods: the improvement increases as the size of the training dataset decreases.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Fine Grained Spoken Document Summarization Through Text Segmentation.\n \n \n \n \n\n\n \n Kotey, S.; Dahyot, R.; and Harte, N.\n\n\n \n\n\n\n In 2022 IEEE Spoken Language Technology Workshop (SLT), pages 647-654, Jan 2023. \n \n\n\n\n
\n\n\n\n \n \n \"FinePaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 6 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@INPROCEEDINGS{KoteySLT2023,\n  author={Kotey, Samantha and Dahyot, Rozenn and Harte, Naomi},\n  booktitle={2022 IEEE Spoken Language Technology Workshop (SLT)}, \n  title={Fine Grained Spoken Document Summarization Through Text Segmentation}, \n  year={2023},\n  volume={},\n  number={},\n  pages={647-654},\n  abstract={Podcast transcripts are long spoken documents of conversational dialogue. Challenging to summarize, podcasts cover a diverse range of topics, vary in length, and have uniquely different linguistic styles. Previous studies in podcast summarization have generated short, concise dialogue summaries. In contrast, we propose a method to generate long fine-grained summaries, which describe details of sub-topic narratives. Leveraging a readability formula, we curate a data subset to train a long sequence transformer for abstractive summarization. Through text segmentation, we filter the evaluation data and exclude specific segments of text. We apply the model to segmented data, producing different types of fine grained summaries. We show that appropriate filtering creates comparable results on ROUGE and serves as an alternative method to truncation. Experiments show our model outperforms previous studies on the Spotify podcast dataset when tasked with generating longer sequences of text.},\n  keywords={},\n  doi={10.1109/SLT54892.2023.10022829},\n  url={https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10022829},\n  ISSN={},\n  month={Jan},}\n\n\n
\n
\n\n\n
\n Podcast transcripts are long spoken documents of conversational dialogue. Challenging to summarize, podcasts cover a diverse range of topics, vary in length, and have uniquely different linguistic styles. Previous studies in podcast summarization have generated short, concise dialogue summaries. In contrast, we propose a method to generate long fine-grained summaries, which describe details of sub-topic narratives. Leveraging a readability formula, we curate a data subset to train a long sequence transformer for abstractive summarization. Through text segmentation, we filter the evaluation data and exclude specific segments of text. We apply the model to segmented data, producing different types of fine grained summaries. We show that appropriate filtering creates comparable results on ROUGE and serves as an alternative method to truncation. Experiments show our model outperforms previous studies on the Spotify podcast dataset when tasked with generating longer sequences of text.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2022\n \n \n (5)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Principal Component Classification.\n \n \n \n \n\n\n \n Dahyot, R.\n\n\n \n\n\n\n Technical Report 2022.\n \n\n\n\n
\n\n\n\n \n \n \"PrincipalPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 9 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@techreport{Dahyot_PCC2022,\n \n  author = {Dahyot, Rozenn},\n  \n  keywords = {Supervised Learning, PCA, classification, metric learning, deep learning, class encoding},\n  abstract={We propose to directly compute classification estimates\nby learning features encoded with their class scores. \nOur resulting model has a encoder-decoder structure suitable for supervised learning, it is computationally efficient and performs well for classification on several datasets.},\n\n  title = {Principal Component Classification},\n  \n  publisher = {arXiv},\n  \n  year = {2022},\n   doi = {10.48550/ARXIV.2210.12746},\n  \n  url = {https://arxiv.org/pdf/2210.12746.pdf},\n  \n  copyright = {Creative Commons Attribution 4.0 International},\n}\n\n
\n
\n\n\n
\n We propose to directly compute classification estimates by learning features encoded with their class scores. Our resulting model has a encoder-decoder structure suitable for supervised learning, it is computationally efficient and performs well for classification on several datasets.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Harmonic Convolutional Networks based on Discrete Cosine Transform.\n \n \n \n \n\n\n \n Ulicny, M.; Krylov, V. A.; and Dahyot, R.\n\n\n \n\n\n\n Pattern Recognition, 129: 1-12. 2022.\n arXiv.2001.06570 Github: https://github.com/matej-ulicny/harmonic-networks\n\n\n\n
\n\n\n\n \n \n \"HarmonicPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 10 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{ULICNY2022108707, \nauthor= {Matej Ulicny and Vladimir A. Krylov and Rozenn Dahyot}, \ntitle= {Harmonic Convolutional Networks based on Discrete Cosine Transform}, \njournal={Pattern Recognition},\nabstract={Convolutional neural networks (CNNs) learn filters in order to capture local correlation patterns in feature space. We propose to learn these filters as combinations of preset spectral filters defined by the Discrete Cosine Transform (DCT). Our proposed DCT-based harmonic blocks replace conventional convolutional layers to produce partially or fully harmonic versions of new or existing CNN architectures. Using DCT energy compaction properties, we demonstrate how the harmonic networks can be efficiently compressed by truncating high-frequency information in harmonic blocks thanks to the redundancies in the spectral domain. We report extensive experimental validation demonstrating benefits of the introduction of harmonic blocks into state-of-the-art CNN models in image classification, object detection and semantic segmentation applications.},\nvolume= {129},\npages={1-12},\nyear= {2022}, \nissn = {0031-3203},\nurl= {https://arxiv.org/pdf/2001.06570.pdf},\ndoi={10.1016/j.patcog.2022.108707},\nnote={arXiv.2001.06570  Github: https://github.com/matej-ulicny/harmonic-networks},\n}\n
\n
\n\n\n
\n Convolutional neural networks (CNNs) learn filters in order to capture local correlation patterns in feature space. We propose to learn these filters as combinations of preset spectral filters defined by the Discrete Cosine Transform (DCT). Our proposed DCT-based harmonic blocks replace conventional convolutional layers to produce partially or fully harmonic versions of new or existing CNN architectures. Using DCT energy compaction properties, we demonstrate how the harmonic networks can be efficiently compressed by truncating high-frequency information in harmonic blocks thanks to the redundancies in the spectral domain. We report extensive experimental validation demonstrating benefits of the introduction of harmonic blocks into state-of-the-art CNN models in image classification, object detection and semantic segmentation applications.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Improving semantic segmentation with graph-based structural knowledge.\n \n \n \n \n\n\n \n Chopin, J.; Fasquel, J.; Mouchere, H.; Dahyot, R.; and Bloch, I.\n\n\n \n\n\n\n In El Yacoubi, M.; Granger, E.; Yuen, P. C.; Pal, U.; and Vincent, N., editor(s), Pattern Recognition and Artificial Intelligence, pages 173–184, Paris, France, June 2022. Springer International Publishing\n hal-03633029\n\n\n\n
\n\n\n\n \n \n \"ImprovingPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 2 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ChopinICPRAI2022a,\ntitle={Improving semantic segmentation with graph-based structural knowledge},\nauthor={J. Chopin and J.-B. Fasquel and H. Mouchere and R. Dahyot and I. Bloch},\nabstract={Deep learning based pipelines for semantic segmentation often\nignore structural information available on annotated images used for\ntraining. We propose a novel post-processing module enforcing structural\nknowledge about the objects of interest to improve segmentation\nresults provided by deep learning. This module corresponds to a “manyto-\none-or-none” inexact graph matching approach, and is formulated as\na quadratic assignment problem. Using two standard measures for evaluation,\nwe show experimentally that our pipeline for segmentation of\n3D MRI data of the brain outperforms the baseline CNN (U-Net) used\nalone. In addition, our approach is shown to be resilient to small training\ndatasets that often limit the performance of deep learning.},\ndoi={10.1007/978-3-031-09037-0_15},\nurl= {https://hal.inria.fr/hal-03633029}, \nnote={hal-03633029},\nbooktitle={Pattern Recognition and Artificial Intelligence},\nyear={2022},\npublisher={Springer International Publishing},\neditor={El Yacoubi, Moun{\\^i}m\nand Granger, Eric\nand Yuen, Pong Chi\nand Pal, Umapada\nand Vincent, Nicole},\nmonth={June},\nHAL_ID = {hal-03633029},\naddress={Paris, France},\nisbn={978-3-031-09037-0},\npages={173--184},\n}
\n
\n\n\n
\n Deep learning based pipelines for semantic segmentation often ignore structural information available on annotated images used for training. We propose a novel post-processing module enforcing structural knowledge about the objects of interest to improve segmentation results provided by deep learning. This module corresponds to a “manyto- one-or-none” inexact graph matching approach, and is formulated as a quadratic assignment problem. Using two standard measures for evaluation, we show experimentally that our pipeline for segmentation of 3D MRI data of the brain outperforms the baseline CNN (U-Net) used alone. In addition, our approach is shown to be resilient to small training datasets that often limit the performance of deep learning.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n QAP Optimisation with Reinforcement Learning for Faster Graph Matching in Sequential Semantic Image Analysis.\n \n \n \n \n\n\n \n Chopin, J.; Fasquel, J.; Mouchere, H.; Dahyot, R.; and Bloch, I.\n\n\n \n\n\n\n In El Yacoubi, M.; Granger, E.; Yuen, P. C.; Pal, U.; and Vincent, N., editor(s), Pattern Recognition and Artificial Intelligence, pages 47–58, Paris, France, June 2022. Springer International Publishing\n hal-03633036\n\n\n\n
\n\n\n\n \n \n \"QAPPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ChopinICPRAI2022b,\ntitle={QAP Optimisation with Reinforcement Learning for Faster Graph Matching in Sequential Semantic Image Analysis},\nauthor={J. Chopin and J.-B. Fasquel and H. Mouchere and R. Dahyot and I. Bloch},\nabstract={The paper addresses the fundamental task of semantic image\nanalysis by exploiting structural information (spatial relationships\nbetween image regions). We propose to perform such semantic image\nanalysis by combining a deep neural network (CNN) with graph matching\nwhere graphs encode efficiently structural information related to regions\nsegmented by the CNN. Our novel approach solves the quadratic assignment\nproblem (QAP) sequentially for matching graphs. The optimal\nsequence for graph matching is conveniently defined using reinforcementlearning\n(RL) based on the region membership probabilities produced by\nthe CNN and their structural relationships. Our RL based strategy for\nsolving QAP sequentially allows us to significantly reduce the combinatioral\ncomplexity for graph matching. Preliminary experiments are performed\non both a synthetic dataset and a public dataset dedicated to the\nsemantic segmentation of face images. Results show that the proposed\nRL-based ordering dramatically outperforms random ordering, and that\nour strategy is about 386 times faster than a global QAP-based approach,\nwhile preserving similar segmentation accuracy.},\npublisher={Springer International Publishing},\neditor={El Yacoubi, Moun{\\^i}m\nand Granger, Eric\nand Yuen, Pong Chi\nand Pal, Umapada\nand Vincent, Nicole},\nisbn={978-3-031-09037-0},\ndoi={10.1007/978-3-031-09037-0_5},\nurl= {https://hal.inria.fr/hal-03633036/}, \nnote={hal-03633036},\nbooktitle={Pattern Recognition and Artificial Intelligence},\nyear={2022},\nmonth={June},\npages={47--58},\naddress={Paris, France},\n}
\n
\n\n\n
\n The paper addresses the fundamental task of semantic image analysis by exploiting structural information (spatial relationships between image regions). We propose to perform such semantic image analysis by combining a deep neural network (CNN) with graph matching where graphs encode efficiently structural information related to regions segmented by the CNN. Our novel approach solves the quadratic assignment problem (QAP) sequentially for matching graphs. The optimal sequence for graph matching is conveniently defined using reinforcementlearning (RL) based on the region membership probabilities produced by the CNN and their structural relationships. Our RL based strategy for solving QAP sequentially allows us to significantly reduce the combinatioral complexity for graph matching. Preliminary experiments are performed on both a synthetic dataset and a public dataset dedicated to the semantic segmentation of face images. Results show that the proposed RL-based ordering dramatically outperforms random ordering, and that our strategy is about 386 times faster than a global QAP-based approach, while preserving similar segmentation accuracy.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n DR-VNet: Retinal Vessel Segmentation via Dense Residual UNet.\n \n \n \n \n\n\n \n Karaali, A.; Dahyot, R.; and Sexton, D. J.\n\n\n \n\n\n\n In El Yacoubi, M.; Granger, E.; Yuen, P. C.; Pal, U.; and Vincent, N., editor(s), Pattern Recognition and Artificial Intelligence, volume abs/2111.04739, Paris, France, June 2022. Springer International Publishing\n Github https://github.com/alikaraali/DR-VNet, ArXivDOI:10.48550/arXiv.2111.04739\n\n\n\n
\n\n\n\n \n \n \"DR-VNet:Paper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 2 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{karaali2022drvnet,\n      title={DR-VNet: Retinal Vessel Segmentation via Dense Residual UNet}, \n      author={Ali Karaali and Rozenn Dahyot and Donal J. Sexton},\n      year={2022},\n\t  booktitle={Pattern Recognition and Artificial Intelligence},\n\t  doi={10.1007/978-3-031-09037-0_17},\n\t  note={Github https://github.com/alikaraali/DR-VNet, ArXivDOI:10.48550/arXiv.2111.04739},\n\t  url= {https://arxiv.org/pdf/2111.04739.pdf}, \n\t  abstract={Accurate retinal vessel segmentation is an important task for many computer-aided diagnosis systems. Yet, it is still a challenging problem due to the complex vessel structures of an eye. Numerous vessel segmentation methods have been proposed recently, however more research is needed to deal with poor segmentation of thin and tiny vessels. To address this, we propose a new deep learning pipeline combining the efficiency of residual dense net blocks and, residual squeeze and excitation blocks. We validate experimentally our approach on three datasets and show that our pipeline outperforms current state of the art techniques on the sensitivity metric relevant to assess capture of small vessels.},\n\t  publisher={Springer International Publishing},\neditor={El Yacoubi, Moun{\\^i}m\nand Granger, Eric\nand Yuen, Pong Chi\nand Pal, Umapada\nand Vincent, Nicole},\nisbn={978-3-031-09037-0},\n      volume= {abs/2111.04739},\n\t  month={June},\naddress={Paris, France},\n      archivePrefix={arXiv},\n      primaryClass={eess.IV}\n}\n
\n
\n\n\n
\n Accurate retinal vessel segmentation is an important task for many computer-aided diagnosis systems. Yet, it is still a challenging problem due to the complex vessel structures of an eye. Numerous vessel segmentation methods have been proposed recently, however more research is needed to deal with poor segmentation of thin and tiny vessels. To address this, we propose a new deep learning pipeline combining the efficiency of residual dense net blocks and, residual squeeze and excitation blocks. We validate experimentally our approach on three datasets and show that our pipeline outperforms current state of the art techniques on the sensitivity metric relevant to assess capture of small vessels.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2021\n \n \n (3)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Context Aware Object Geotagging.\n \n \n \n \n\n\n \n Liu, C.; Ulicny, M.; Manzke, M.; and Dahyot, R.\n\n\n \n\n\n\n In Irish Machine Vision and Image Processing (IMVIP 2021), 2021. \n \n\n\n\n
\n\n\n\n \n \n \"ContextPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 7 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ChaoImvip2021,\nauthor= {C.-J. Liu and Matej Ulicny and Michael Manzke and  Rozenn Dahyot}, \ntitle= {Context Aware Object Geotagging},\nbooktitle= {Irish Machine Vision and Image Processing (IMVIP 2021)},\nvolume= {},\nyear= {2021},\nabstract={We propose an approach for geolocating assets from street view imagery \nby improving the quality of the metadata associated with the images using \nStructure from Motion, and by using contextual geographic information extracted \nfrom OpenStreetMap. Our pipeline is validated experimentally against the state of\n the art approaches for geotagging traffic lights.},\nurl= {https://arxiv.org/pdf/2108.06302.pdf},\ndoi={10.48550/arXiv.2108.06302},\nnote={},\narchivePrefix= {arXiv}, \neprint= {},\ntimestamp= {},\nbiburl= {},\nbibsource= {}\n}
\n
\n\n\n
\n We propose an approach for geolocating assets from street view imagery by improving the quality of the metadata associated with the images using Structure from Motion, and by using contextual geographic information extracted from OpenStreetMap. Our pipeline is validated experimentally against the state of the art approaches for geotagging traffic lights.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Model for predicting perception of facial action unit activation using virtual humans.\n \n \n \n \n\n\n \n McDonnell, R.; Zibrek, K.; Carrigan, E.; and Dahyot, R.\n\n\n \n\n\n\n Computers & Graphics , 100: 81-92. 2021.\n Winner 2022 Graphics Replicability Stamp Initiative (GRSI) best paper award; Github: https://github.com/Roznn/facial-blendshapes\n\n\n\n
\n\n\n\n \n \n \"ModelPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 5 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n\n\n\n
\n
@article{McDonnell2021,\n title= {Model for predicting perception of facial action unit activation using virtual humans},\n journal= {Computers \\& Graphics }, \ndoi = {10.1016/j.cag.2021.07.022},\n volume= {100}, \n pages= {81-92}, \n year= {2021}, \n note= {Winner 2022 Graphics Replicability Stamp Initiative (GRSI) best paper award; Github: https://github.com/Roznn/facial-blendshapes}, \n issn= {0097-8493},\n url= {https://roznn.github.io/facial-blendshapes/CAG2021.pdf}, \n author= {Rachel McDonnell and Katja Zibrek and Emma Carrigan and Rozenn Dahyot}, \n keywords= {facial action unit, perception, virtual character},\n abstract= {Blendshape facial rigs are used extensively in the industry for facial animation of\nvirtual humans. However, storing and manipulating large numbers of facial meshes\n(blendshapes) is costly in terms of memory and computation for gaming applications.\nBlendshape rigs are comprised of sets of semantically-meaningful expressions, which\ngovern how expressive the character will be, often based on Action Units from the Facial\nAction Coding System (FACS). However, the relative perceptual importance of blendshapes has not yet been investigated. Research in Psychology and Neuroscience has\nshown that our brains process faces differently than other objects so we postulate that\nthe perception of facial expressions will be feature-dependent rather than based purely\non the amount of movement required to make the expression. Therefore, we believe that\nperception of blendshape visibility will not be reliably predicted by numerical calculations of the difference between the expression and the neutral mesh. In this paper, we\nexplore the noticeability of blendshapes under different activation levels, and present\nnew perceptually-based models to predict perceptual importance of blendshapes. The\nmodels predict visibility based on commonly-used geometry and image-based metrics.}\n }
\n
\n\n\n
\n Blendshape facial rigs are used extensively in the industry for facial animation of virtual humans. However, storing and manipulating large numbers of facial meshes (blendshapes) is costly in terms of memory and computation for gaming applications. Blendshape rigs are comprised of sets of semantically-meaningful expressions, which govern how expressive the character will be, often based on Action Units from the Facial Action Coding System (FACS). However, the relative perceptual importance of blendshapes has not yet been investigated. Research in Psychology and Neuroscience has shown that our brains process faces differently than other objects so we postulate that the perception of facial expressions will be feature-dependent rather than based purely on the amount of movement required to make the expression. Therefore, we believe that perception of blendshape visibility will not be reliably predicted by numerical calculations of the difference between the expression and the neutral mesh. In this paper, we explore the noticeability of blendshapes under different activation levels, and present new perceptually-based models to predict perceptual importance of blendshapes. The models predict visibility based on commonly-used geometry and image-based metrics.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Sliced L2 Distance for Colour Grading.\n \n \n \n \n\n\n \n Alghamdi, H.; and Dahyot, R.\n\n\n \n\n\n\n In 2021 29th European Signal Processing Conference (EUSIPCO), pages 671-675, 2021. \n https://arxiv.org/pdf/2102.09297.pdf\n\n\n\n
\n\n\n\n \n \n \"SlicedPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 6 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{alghamdi2021sliced,\n      title = {Sliced L2 Distance for Colour Grading}, \n      author = {Hana Alghamdi and Rozenn Dahyot},\n\t  booktitle = {2021 29th European Signal Processing Conference (EUSIPCO)},\n\t  doi = {10.23919/EUSIPCO54536.2021.9616260},\n      year = {2021},\n\t  volume={},\n      number={},\n      pages={671-675},\n      eprint = {2102.09297},\n\t  archivePrefix = {arXiv},\n      primaryClass = {cs.CV},\n\t  abstract = {We propose a new method with L2 distance that maps one N-dimensional distribution to another,\n\t  taking into account available information about correspondences. We solve the high-dimensional problem \n\t  in 1D space using an iterative projection approach. To show the potentials of this mapping, we apply it\n\t  to colour transfer between two images that exhibit overlapped scenes. Experiments show quantitative and \n\t  qualitative competitive results as compared with the state of the art colour transfer methods.},\n\t  note={https://arxiv.org/pdf/2102.09297.pdf},\n\t  url = {https://eurasip.org/Proceedings/Eusipco/Eusipco2021/pdfs/0000671.pdf}\n}
\n
\n\n\n
\n We propose a new method with L2 distance that maps one N-dimensional distribution to another, taking into account available information about correspondences. We solve the high-dimensional problem in 1D space using an iterative projection approach. To show the potentials of this mapping, we apply it to colour transfer between two images that exhibit overlapped scenes. Experiments show quantitative and qualitative competitive results as compared with the state of the art colour transfer methods.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n\n\n\n
\n\n\n \n\n \n \n \n \n\n
\n"}; document.write(bibbase_data.data);