var bibbase_data = {"data":"\"Loading..\"\n\n
\n\n \n\n \n\n \n \n\n \n\n \n \n\n \n\n \n
\n generated by\n \n \"bibbase.org\"\n\n \n
\n \n\n
\n\n \n\n\n
\n\n Excellent! Next you can\n create a new website with this list, or\n embed it in an existing web page by copying & pasting\n any of the following snippets.\n\n
\n JavaScript\n (easiest)\n
\n \n <script src=\"https://bibbase.org/show?bib=https://florentfo.rest/files/publications.bib&jsonp=1&css=1&commas=1&fullnames=1&jsonp=1\"></script>\n \n
\n\n PHP\n
\n \n <?php\n $contents = file_get_contents(\"https://bibbase.org/show?bib=https://florentfo.rest/files/publications.bib&jsonp=1&css=1&commas=1&fullnames=1\");\n print_r($contents);\n ?>\n \n
\n\n iFrame\n (not recommended)\n
\n \n <iframe src=\"https://bibbase.org/show?bib=https://florentfo.rest/files/publications.bib&jsonp=1&css=1&commas=1&fullnames=1\"></iframe>\n \n
\n\n

\n For more details see the documention.\n

\n
\n
\n\n
\n\n This is a preview! To use this list on your own web site\n or create a new web site from it,\n create a free account. The file will be added\n and you will be able to edit it in the File Manager.\n We will show you instructions once you've created your account.\n
\n\n
\n\n

To the site owner:

\n\n

Action required! Mendeley is changing its\n API. In order to keep using Mendeley with BibBase past April\n 14th, you need to:\n

    \n
  1. renew the authorization for BibBase on Mendeley, and
  2. \n
  3. update the BibBase URL\n in your page the same way you did when you initially set up\n this page.\n
  4. \n
\n

\n\n

\n \n \n Fix it now\n

\n
\n\n
\n\n\n
\n \n \n
\n
\n  \n 2024\n \n \n (3)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Health Prediction for Lithium-Ion Batteries Under Unseen Working Conditions.\n \n \n \n \n\n\n \n Yunhong Che, Florent Forest, Yusheng Zheng, Le Xu, & Remus Teodorescu.\n\n\n \n\n\n\n IEEE Transactions on Industrial Electronics,1–11. april 2024.\n \n\n\n\n
\n\n\n\n \n \n \"Health link\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@article{che2024health,\n\ttitle = {Health {Prediction} for {Lithium}-{Ion} {Batteries} {Under} {Unseen} {Working} {Conditions}},\n\tissn = {1557-9948},\n\tdoi = {10.1109/TIE.2024.3379664},\n\tabstract = {Battery health prediction is significant while challenging for intelligent battery management. This article proposes a general framework for both short-term and long-term predictions of battery health under unseen dynamic loading and temperature conditions using domain-adaptive multitask learning (MTL) with long-term regularization. First, features extracted from partial charging curves are utilized for short-term state of health predictions. Then, the long-term degradation trajectory is directly predicted by recursively using the predicted features within the multitask framework, enhancing the model integrity and lowering the complexity. Then, domain adaptation (DA) is adopted to reduce the discrepancies between different working conditions. Additionally, a long-term regularization is introduced to address the shortcoming that arises when the model is extrapolated recursively for future health predictions. Thus, the short-term prediction ability is maintained while the long-term prediction performance is enhanced. Finally, predictions are validated through aging experiments under various dynamic loading profiles. By using partial charging capacity–voltage data, the results show that the early-stage long-term predictions are accurate and stable under various working profiles, with root mean square errors below 2\\% and fitting coefficients surpassing 0.86.},\n\tjournal = {IEEE Transactions on Industrial Electronics},\n\tauthor = {Che, Yunhong and Forest, Florent and Zheng, Yusheng and Xu, Le and Teodorescu, Remus},\n\tyear = {2024},\n\tmonth = april,\n\tkeywords = {Aging, Batteries, Battery, Degradation, domain adaptation (DA), Feature extraction, health and trajectory prediction, Loading, multi-task learning, Predictive models, Testing, transfer learning},\n\tpages = {1--11},\n\turl_Link = {https://ieeexplore.ieee.org/document/10500447},\n\tbibbase_note = {<img src="assets/img/papers/battery-conditions.png">}\n}\n\n
\n
\n\n\n
\n Battery health prediction is significant while challenging for intelligent battery management. This article proposes a general framework for both short-term and long-term predictions of battery health under unseen dynamic loading and temperature conditions using domain-adaptive multitask learning (MTL) with long-term regularization. First, features extracted from partial charging curves are utilized for short-term state of health predictions. Then, the long-term degradation trajectory is directly predicted by recursively using the predicted features within the multitask framework, enhancing the model integrity and lowering the complexity. Then, domain adaptation (DA) is adopted to reduce the discrepancies between different working conditions. Additionally, a long-term regularization is introduced to address the shortcoming that arises when the model is extrapolated recursively for future health predictions. Thus, the short-term prediction ability is maintained while the long-term prediction performance is enhanced. Finally, predictions are validated through aging experiments under various dynamic loading profiles. By using partial charging capacity–voltage data, the results show that the early-stage long-term predictions are accurate and stable under various working profiles, with root mean square errors below 2% and fitting coefficients surpassing 0.86.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n ThermoNeRF: Multimodal Neural Radiance Fields for Thermal Novel View Synthesis.\n \n \n \n \n\n\n \n Mariam Hassan, Florent Forest, Olga Fink, & Malcolm Mielle.\n\n\n \n\n\n\n March 2024.\n arXiv:2403.12154 [cs]\n\n\n\n
\n\n\n\n \n \n \"ThermoNeRF: link\n  \n \n \n \"ThermoNeRF: paper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@misc{hassan2024thermonerf,\n\ttitle = {{ThermoNeRF}: {Multimodal} {Neural} {Radiance} {Fields} for {Thermal} {Novel} {View} {Synthesis}},\n\tcopyright = {All rights reserved},\n\tshorttitle = {{ThermoNeRF}},\n\tdoi = {10.48550/arXiv.2403.12154},\n\tabstract = {Thermal scene reconstruction exhibit great potential for applications across a broad spectrum of fields, including building energy consumption analysis and non-destructive testing. However, existing methods typically require dense scene measurements and often rely on RGB images for 3D geometry reconstruction, with thermal information being projected post-reconstruction. This two-step strategy, adopted due to the lack of texture in thermal images, can lead to disparities between the geometry and temperatures of the reconstructed objects and those of the actual scene. To address this challenge, we propose ThermoNeRF, a novel multimodal approach based on Neural Radiance Fields, capable of rendering new RGB and thermal views of a scene jointly. To overcome the lack of texture in thermal images, we use paired RGB and thermal images to learn scene density, while distinct networks estimate color and temperature information. Furthermore, we introduce ThermoScenes, a new dataset to palliate the lack of available RGB+thermal datasets for scene reconstruction. Experimental results validate that ThermoNeRF achieves accurate thermal image synthesis, with an average mean absolute error of 1.5$\\circ$C, an improvement of over 50\\% compared to using concatenated RGB+thermal data with Nerfacto, a state-of-the-art NeRF method.},\n\turldate = {2024-03-20},\n\tpublisher = {arXiv},\n\tauthor = {Hassan, Mariam and Forest, Florent and Fink, Olga and Mielle, Malcolm},\n\tmonth = mar,\n\tyear = {2024},\n\tnote = {arXiv:2403.12154 [cs]},\n\turl_Link = {http://arxiv.org/abs/2403.12154},\n\turl_Paper = {http://arxiv.org/pdf/2403.12154.pdf},\n\tbibbase_note = {<img src="assets/img/papers/thermonerf.png">}\n}\n\n
\n
\n\n\n
\n Thermal scene reconstruction exhibit great potential for applications across a broad spectrum of fields, including building energy consumption analysis and non-destructive testing. However, existing methods typically require dense scene measurements and often rely on RGB images for 3D geometry reconstruction, with thermal information being projected post-reconstruction. This two-step strategy, adopted due to the lack of texture in thermal images, can lead to disparities between the geometry and temperatures of the reconstructed objects and those of the actual scene. To address this challenge, we propose ThermoNeRF, a novel multimodal approach based on Neural Radiance Fields, capable of rendering new RGB and thermal views of a scene jointly. To overcome the lack of texture in thermal images, we use paired RGB and thermal images to learn scene density, while distinct networks estimate color and temperature information. Furthermore, we introduce ThermoScenes, a new dataset to palliate the lack of available RGB+thermal datasets for scene reconstruction. Experimental results validate that ThermoNeRF achieves accurate thermal image synthesis, with an average mean absolute error of 1.5$∘$C, an improvement of over 50% compared to using concatenated RGB+thermal data with Nerfacto, a state-of-the-art NeRF method.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Uncertainty-Guided Alignment for Unsupervised Domain Adaptation in Regression.\n \n \n \n \n\n\n \n Ismail Nejjar, Gaëtan Frusque, Florent Forest, & Olga Fink.\n\n\n \n\n\n\n January 2024.\n arXiv:2401.13721 [cs]\n\n\n\n
\n\n\n\n \n \n \"Uncertainty-Guided link\n  \n \n \n \"Uncertainty-Guided paper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@misc{nejjar2024uncertainty,\n\ttitle = {Uncertainty-{Guided} {Alignment} for {Unsupervised} {Domain} {Adaptation} in {Regression}},\n\tcopyright = {All rights reserved},\n\tdoi = {10.48550/arXiv.2401.13721},\n\tabstract = {Unsupervised Domain Adaptation for Regression (UDAR) aims to adapt a model from a labeled source domain to an unlabeled target domain for regression tasks. Recent successful works in UDAR mostly focus on subspace alignment, involving the alignment of a selected subspace within the entire feature space. This contrasts with the feature alignment methods used for classification, which aim at aligning the entire feature space and have proven effective but are less so in regression settings. Specifically, while classification aims to identify separate clusters across the entire embedding dimension, regression induces less structure in the data representation, necessitating additional guidance for efficient alignment. In this paper, we propose an effective method for UDAR by incorporating guidance from uncertainty. Our approach serves a dual purpose: providing a measure of confidence in predictions and acting as a regularization of the embedding space. Specifically, we leverage the Deep Evidential Learning framework, which outputs both predictions and uncertainties for each input sample. We propose aligning the parameters of higher-order evidential distributions between the source and target domains using traditional alignment methods at the feature or posterior level. Additionally, we propose to augment the feature space representation by mixing source samples with pseudo-labeled target samples based on label similarity. This cross-domain mixing strategy produces more realistic samples than random mixing and introduces higher uncertainty, facilitating further alignment. We demonstrate the effectiveness of our approach on four benchmarks for UDAR, on which we outperform existing methods.},\n\turldate = {2024-03-21},\n\tpublisher = {arXiv},\n\tauthor = {Nejjar, Ismail and Frusque, Gaëtan and Forest, Florent and Fink, Olga},\n\tmonth = jan,\n\tyear = {2024},\n\tnote = {arXiv:2401.13721 [cs]},\n\turl_Link = {http://arxiv.org/abs/2401.13721},\n\turl_Paper = {http://arxiv.org/pdf/2401.13721.pdf},\n\tbibbase_note = {<img src="assets/img/papers/uga.png">}\n}\n\n
\n
\n\n\n
\n Unsupervised Domain Adaptation for Regression (UDAR) aims to adapt a model from a labeled source domain to an unlabeled target domain for regression tasks. Recent successful works in UDAR mostly focus on subspace alignment, involving the alignment of a selected subspace within the entire feature space. This contrasts with the feature alignment methods used for classification, which aim at aligning the entire feature space and have proven effective but are less so in regression settings. Specifically, while classification aims to identify separate clusters across the entire embedding dimension, regression induces less structure in the data representation, necessitating additional guidance for efficient alignment. In this paper, we propose an effective method for UDAR by incorporating guidance from uncertainty. Our approach serves a dual purpose: providing a measure of confidence in predictions and acting as a regularization of the embedding space. Specifically, we leverage the Deep Evidential Learning framework, which outputs both predictions and uncertainties for each input sample. We propose aligning the parameters of higher-order evidential distributions between the source and target domains using traditional alignment methods at the feature or posterior level. Additionally, we propose to augment the feature space representation by mixing source samples with pseudo-labeled target samples based on label similarity. This cross-domain mixing strategy produces more realistic samples than random mixing and introduces higher uncertainty, facilitating further alignment. We demonstrate the effectiveness of our approach on four benchmarks for UDAR, on which we outperform existing methods.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2023\n \n \n (6)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Calibrated Adaptive Teacher for Domain Adaptive Intelligent Fault Diagnosis.\n \n \n \n \n\n\n \n Florent Forest, & Olga Fink.\n\n\n \n\n\n\n December 2023.\n arXiv:2312.02826 [cs, eess, stat]\n\n\n\n
\n\n\n\n \n \n \"Calibrated link\n  \n \n \n \"Calibrated paper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@misc{forest2023calibratedadaptive,\n\ttitle = {Calibrated {Adaptive} {Teacher} for {Domain} {Adaptive} {Intelligent} {Fault} {Diagnosis}},\n\tcopyright = {All rights reserved},\n\tdoi = {10.48550/arXiv.2312.02826},\n\tabstract = {Intelligent Fault Diagnosis (IFD) based on deep learning has proven to be an effective and flexible solution, attracting extensive research. Deep neural networks can learn rich representations from vast amounts of representative labeled data for various applications. In IFD, they achieve high classification performance from signals in an end-to-end manner, without requiring extensive domain knowledge. However, deep learning models usually only perform well on the data distribution they have been trained on. When applied to a different distribution, they may experience performance drops. This is also observed in IFD, where assets are often operated in working conditions different from those in which labeled data have been collected. Unsupervised domain adaptation (UDA) deals with the scenario where labeled data are available in a source domain, and only unlabeled data are available in a target domain, where domains may correspond to operating conditions. Recent methods rely on training with confident pseudo-labels for target samples. However, the confidence-based selection of pseudo-labels is hindered by poorly calibrated confidence estimates in the target domain, primarily due to over-confident predictions, which limits the quality of pseudo-labels and leads to error accumulation. In this paper, we propose a novel UDA method called Calibrated Adaptive Teacher (CAT), where we propose to calibrate the predictions of the teacher network throughout the self-training process, leveraging post-hoc calibration techniques. We evaluate CAT on domain-adaptive IFD and perform extensive experiments on the Paderborn benchmark for bearing fault diagnosis under varying operating conditions. Our proposed method achieves state-of-the-art performance on most transfer tasks.},\n\tpublisher = {arXiv},\n\tauthor = {Forest, Florent and Fink, Olga},\n\tmonth = dec,\n\tyear = {2023},\n\tnote = {arXiv:2312.02826 [cs, eess, stat]},\n    url_Link = {http://arxiv.org/abs/2312.02826},\n    url_Paper = {http://arxiv.org/pdf/2312.02826.pdf},\n\tbibbase_note = {<img src="assets/img/papers/cat.png">}\n}\n\n
\n
\n\n\n
\n Intelligent Fault Diagnosis (IFD) based on deep learning has proven to be an effective and flexible solution, attracting extensive research. Deep neural networks can learn rich representations from vast amounts of representative labeled data for various applications. In IFD, they achieve high classification performance from signals in an end-to-end manner, without requiring extensive domain knowledge. However, deep learning models usually only perform well on the data distribution they have been trained on. When applied to a different distribution, they may experience performance drops. This is also observed in IFD, where assets are often operated in working conditions different from those in which labeled data have been collected. Unsupervised domain adaptation (UDA) deals with the scenario where labeled data are available in a source domain, and only unlabeled data are available in a target domain, where domains may correspond to operating conditions. Recent methods rely on training with confident pseudo-labels for target samples. However, the confidence-based selection of pseudo-labels is hindered by poorly calibrated confidence estimates in the target domain, primarily due to over-confident predictions, which limits the quality of pseudo-labels and leads to error accumulation. In this paper, we propose a novel UDA method called Calibrated Adaptive Teacher (CAT), where we propose to calibrate the predictions of the teacher network throughout the self-training process, leveraging post-hoc calibration techniques. We evaluate CAT on domain-adaptive IFD and perform extensive experiments on the Paderborn benchmark for bearing fault diagnosis under varying operating conditions. Our proposed method achieves state-of-the-art performance on most transfer tasks.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n From Classification to Segmentation with Explainable AI: A Study on Crack Detection and Growth Monitoring.\n \n \n \n \n\n\n \n Florent Forest, Hugo Porta, Devis Tuia, & Olga Fink.\n\n\n \n\n\n\n September 2023.\n arXiv:2309.11267 [cs, eess]\n\n\n\n
\n\n\n\n \n \n \"From link\n  \n \n \n \"From paper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@misc{forest2023classification,\n\ttitle = {From {Classification} to {Segmentation} with {Explainable} {AI}: {A} {Study} on {Crack} {Detection} and {Growth} {Monitoring}},\n\tcopyright = {All rights reserved},\n\tshorttitle = {From {Classification} to {Segmentation} with {Explainable} {AI}},\n\tdoi = {10.48550/arXiv.2309.11267},\n\tabstract = {Monitoring surface cracks in infrastructure is crucial for structural health monitoring. Automatic visual inspection offers an effective solution, especially in hard-to-reach areas. Machine learning approaches have proven their effectiveness but typically require large annotated datasets for supervised training. Once a crack is detected, monitoring its severity often demands precise segmentation of the damage. However, pixel-level annotation of images for segmentation is labor-intensive. To mitigate this cost, one can leverage explainable artificial intelligence (XAI) to derive segmentations from the explanations of a classifier, requiring only weak image-level supervision. This paper proposes applying this methodology to segment and monitor surface cracks. We evaluate the performance of various XAI methods and examine how this approach facilitates severity quantification and growth monitoring. Results reveal that while the resulting segmentation masks may exhibit lower quality than those produced by supervised methods, they remain meaningful and enable severity monitoring, thus reducing substantial labeling costs.},\n\tpublisher = {arXiv},\n\tauthor = {Forest, Florent and Porta, Hugo and Tuia, Devis and Fink, Olga},\n\tmonth = sep,\n\tyear = {2023},\n\tnote = {arXiv:2309.11267 [cs, eess]},\n\turl_Link = {http://arxiv.org/abs/2309.11267},\n\turl_Paper = {http://arxiv.org/pdf/2309.11267.pdf},\n\tbibbase_note = {<img src="assets/img/papers/crack-explain.png">}\n}\n\n
\n
\n\n\n
\n Monitoring surface cracks in infrastructure is crucial for structural health monitoring. Automatic visual inspection offers an effective solution, especially in hard-to-reach areas. Machine learning approaches have proven their effectiveness but typically require large annotated datasets for supervised training. Once a crack is detected, monitoring its severity often demands precise segmentation of the damage. However, pixel-level annotation of images for segmentation is labor-intensive. To mitigate this cost, one can leverage explainable artificial intelligence (XAI) to derive segmentations from the explanations of a classifier, requiring only weak image-level supervision. This paper proposes applying this methodology to segment and monitor surface cracks. We evaluate the performance of various XAI methods and examine how this approach facilitates severity quantification and growth monitoring. Results reveal that while the resulting segmentation masks may exhibit lower quality than those produced by supervised methods, they remain meaningful and enable severity monitoring, thus reducing substantial labeling costs.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Calibrated Self-Training for Cross-Domain Bearing Fault Diagnosis.\n \n \n \n \n\n\n \n Florent Forest, & Olga Fink.\n\n\n \n\n\n\n In Proceedings of the 33rd European Safety and Reliability Conference, pages 3406–3407, 2023. \n \n\n\n\n
\n\n\n\n \n \n \"Calibrated link\n  \n \n \n \"Calibrated paper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{forest2023calibratedself,\n\ttitle = {Calibrated {Self}-{Training} for {Cross}-{Domain} {Bearing} {Fault} {Diagnosis}},\n\tcopyright = {All rights reserved},\n\tisbn = {978-981-18807-1-1},\n\tdoi = {10.3850/978-981-18-8071-1_P249-cd},\n\tabstract = {Fault diagnosis of rolling bearings is a crucial task in Prognostics and Health Management, as rolling elements are ubiquitous in industrial assets. Data-driven approaches based on deep neural networks have made significant progress in this area. However, they require collecting large representative labeled data sets. However, in industrial settings, assets are often operated in conditions different from the ones in which labeled data were collected, requiring a transfer between working conditions. In this work, we tackle the classification of bearing fault types and severity levels in the setting of unsupervised domain adaptation (UDA), where labeled data are available in a source domain and only unlabeled data are available in a different but related target domain. We focus on UDA with self-training methods, based on pseudo-labeling of target samples. One major challenge in these methods is to avoid error accumulation due to low-quality pseudo-labels. To address this challenge, we propose incorporating post-hoc calibration, such as the well-known temperature scaling, into the self-training process to increase the quality of selected pseudo-labels. We implement our proposed calibration approach in two self-training algorithms, Calibrated Pseudo-Labeling and Calibrated Adaptive Teacher, and demonstrate their competitive results on the Paderborn University (PU) benchmark for fault diagnosis of rolling bearings under varying operating conditions.},\n\tlanguage = {en},\n\tbooktitle = {Proceedings of the 33rd {European} {Safety} and {Reliability} {Conference}},\n\tauthor = {Forest, Florent and Fink, Olga},\n\tyear = {2023},\n\tpages = {3406--3407},\n\turl_Link = {https://www.rpsonline.com.sg/proceedings/esrel2023/html/P249.html},\n\turl_Paper = {https://www.rpsonline.com.sg/proceedings/esrel2023/pdf/P249.pdf},\n}\n\n
\n
\n\n\n
\n Fault diagnosis of rolling bearings is a crucial task in Prognostics and Health Management, as rolling elements are ubiquitous in industrial assets. Data-driven approaches based on deep neural networks have made significant progress in this area. However, they require collecting large representative labeled data sets. However, in industrial settings, assets are often operated in conditions different from the ones in which labeled data were collected, requiring a transfer between working conditions. In this work, we tackle the classification of bearing fault types and severity levels in the setting of unsupervised domain adaptation (UDA), where labeled data are available in a source domain and only unlabeled data are available in a different but related target domain. We focus on UDA with self-training methods, based on pseudo-labeling of target samples. One major challenge in these methods is to avoid error accumulation due to low-quality pseudo-labels. To address this challenge, we propose incorporating post-hoc calibration, such as the well-known temperature scaling, into the self-training process to increase the quality of selected pseudo-labels. We implement our proposed calibration approach in two self-training algorithms, Calibrated Pseudo-Labeling and Calibrated Adaptive Teacher, and demonstrate their competitive results on the Paderborn University (PU) benchmark for fault diagnosis of rolling bearings under varying operating conditions.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Segmenting Without Annotating: Crack Segmentation and Monitoring via Post-Hoc Classifier Explanations.\n \n \n \n \n\n\n \n Florent Forest, Hugo Porta, Devis Tuia, & Olga Fink.\n\n\n \n\n\n\n In Proceedings of the 33rd European Safety and Reliability Conference, pages 1392–1393, 2023. \n \n\n\n\n
\n\n\n\n \n \n \"Segmenting link\n  \n \n \n \"Segmenting paper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{forest2023segmenting,\n\ttitle = {Segmenting {Without} {Annotating}: {Crack} {Segmentation} and {Monitoring} via {Post}-{Hoc} {Classifier} {Explanations}},\n\tcopyright = {All rights reserved},\n\tisbn = {978-981-18807-1-1},\n\tshorttitle = {Segmenting {Without} {Annotating}},\n\tdoi = {10.3850/978-981-18-8071-1_P290-cd},\n\tabstract = {Monitoring the cracks in walls, roads and other types of infrastructure is essential to ensure the safety of a structure, and plays an important role in structural health monitoring. Automatic visual inspection allows an efficient, costeffective and safe health monitoring, especially in hard-to-reach locations. To this aim, data-driven approaches based on machine learning have demonstrated their effectiveness, at the expense of annotating large sets of images for supervised training. Once a damage has been detected, one also needs to monitor the evolution of its severity, in order to trigger a timely maintenance operation and avoid any catastrophic consequence. This evaluation requires a precise segmentation of the damage. However, pixel-level annotation of images for segmentation is labor-intensive. On the other hand, labeling images for a classification task is relatively cheap in comparison. To circumvent the cost of annotating images for segmentation, recent works inspired by explainable AI (XAI) have proposed to use the post-hoc explanations of a classifier to obtain a segmentation of the input image. In this work, we study the application of XAI techniques to the detection and monitoring of cracks in masonry wall surfaces. We benchmark different post-hoc explainability methods in terms of segmentation quality and accuracy of the damage severity quantification (for example, the width of a crack), thus enabling timely decision-making.},\n\tlanguage = {en},\n\tbooktitle = {Proceedings of the 33rd {European} {Safety} and {Reliability} {Conference}},\n\tauthor = {Forest, Florent and Porta, Hugo and Tuia, Devis and Fink, Olga},\n\tyear = {2023},\n\tpages = {1392--1393},\n\turl_Link = {https://www.rpsonline.com.sg/proceedings/esrel2023/html/P290.html},\n\turl_Paper = {https://www.rpsonline.com.sg/proceedings/esrel2023/pdf/P290.pdf},\n}\n\n
\n
\n\n\n
\n Monitoring the cracks in walls, roads and other types of infrastructure is essential to ensure the safety of a structure, and plays an important role in structural health monitoring. Automatic visual inspection allows an efficient, costeffective and safe health monitoring, especially in hard-to-reach locations. To this aim, data-driven approaches based on machine learning have demonstrated their effectiveness, at the expense of annotating large sets of images for supervised training. Once a damage has been detected, one also needs to monitor the evolution of its severity, in order to trigger a timely maintenance operation and avoid any catastrophic consequence. This evaluation requires a precise segmentation of the damage. However, pixel-level annotation of images for segmentation is labor-intensive. On the other hand, labeling images for a classification task is relatively cheap in comparison. To circumvent the cost of annotating images for segmentation, recent works inspired by explainable AI (XAI) have proposed to use the post-hoc explanations of a classifier to obtain a segmentation of the input image. In this work, we study the application of XAI techniques to the detection and monitoring of cracks in masonry wall surfaces. We benchmark different post-hoc explainability methods in terms of segmentation quality and accuracy of the damage severity quantification (for example, the width of a crack), thus enabling timely decision-making.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Predictive Health Assessment for Lithium-ion Batteries with Probabilistic Degradation Prediction and Accelerating Aging Detection.\n \n \n \n \n\n\n \n Yunhong Che, Yusheng Zheng, Florent Forest, Xin Sui, Xiaosong Hu, & Remus Teodorescu.\n\n\n \n\n\n\n Reliability Engineering & System Safety,109603. August 2023.\n \n\n\n\n
\n\n\n\n \n \n \"Predictive link\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@article{che2023predictive,\n    title = {Predictive {Health} {Assessment} for {Lithium}-ion {Batteries} with {Probabilistic} {Degradation} {Prediction} and {Accelerating} {Aging} {Detection}},\n    copyright = {All rights reserved},\n    issn = {0951-8320},\n    doi = {10.1016/j.ress.2023.109603},\n    abstract = {Predictive health assessment is of vital importance for smarter battery management to ensure optimal and safe operations and thus make the most use of battery life. This paper proposes a general framework for battery aging prognostics in order to provide the predictions of battery knee, lifetime, state of health degradation, and aging rate variations, as well as the assessment of battery health. Early information is used to predict knee slope and other life-related information via deep multi-task learning, where the convolutional-long-short-term memory-bayesian neural network is proposed. The structure is also used for online state of health and degradation rate predictions for the detection of accelerating aging. The two probabilistic predicted boundaries identify the accelerating aging regions for battery health assessment. To avoid wrong and premature alarms, the empirical model is used for data preprocessing and the slope is predicted together with the state of health via multi-task learning. A cloud-edge framework is considered where fine-tuning is adopted for performance improvement during cycling. The proposed general framework is flexible for adjustment to different practical requirements and can be extrapolated to other batteries aged under different conditions. The results indicate that the early predictions are improved using the proposed method compared to multiple single feature-based benchmarks, and that integration of the algorithm is improved. The sequence prediction is reliable for different predicted lengths with root mean square errors of less than 1.41\\%, and the detection of accelerating aging can guide reliable predictive health management.},\n    journal = {Reliability Engineering \\& System Safety},\n    author = {Che, Yunhong and Zheng, Yusheng and Forest, Florent and Sui, Xin and Hu, Xiaosong and Teodorescu, Remus},\n    month = aug,\n    year = {2023},\n    keywords = {Transfer learning, Battery degradation prediction, Knee point detection, Multi-task learning, Predictive health assessment, Probabilistic prediction},\n    pages = {109603},\n    url_Link = {https://www.sciencedirect.com/science/article/pii/S0951832023005173},\n    bibbase_note = {<img src="assets/img/papers/battery-aging.png">}\n}\n\n
\n
\n\n\n
\n Predictive health assessment is of vital importance for smarter battery management to ensure optimal and safe operations and thus make the most use of battery life. This paper proposes a general framework for battery aging prognostics in order to provide the predictions of battery knee, lifetime, state of health degradation, and aging rate variations, as well as the assessment of battery health. Early information is used to predict knee slope and other life-related information via deep multi-task learning, where the convolutional-long-short-term memory-bayesian neural network is proposed. The structure is also used for online state of health and degradation rate predictions for the detection of accelerating aging. The two probabilistic predicted boundaries identify the accelerating aging regions for battery health assessment. To avoid wrong and premature alarms, the empirical model is used for data preprocessing and the slope is predicted together with the state of health via multi-task learning. A cloud-edge framework is considered where fine-tuning is adopted for performance improvement during cycling. The proposed general framework is flexible for adjustment to different practical requirements and can be extrapolated to other batteries aged under different conditions. The results indicate that the early predictions are improved using the proposed method compared to multiple single feature-based benchmarks, and that integration of the algorithm is improved. The sequence prediction is reliable for different predicted lengths with root mean square errors of less than 1.41%, and the detection of accelerating aging can guide reliable predictive health management.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Selecting the Number of Clusters $K$ with a Stability Trade-off: an Internal Validation Criterion.\n \n \n \n \n\n\n \n Alex Mourer, Florent Forest, Mustapha Lebbah, Hanane Azzag, & Jérôme Lacaille.\n\n\n \n\n\n\n In PAKDD, Osaka, Japan, May 2023. \n arXiv:2006.08530 [cs, stat]\n\n\n\n
\n\n\n\n \n \n \"Selecting link\n  \n \n \n \"Selecting paper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{mourer2023selecting,\n    address = {Osaka, Japan},\n    title = {Selecting the {Number} of {Clusters} $K$ with a {Stability} {Trade}-off: an {Internal} {Validation} {Criterion}},\n    copyright = {All rights reserved},\n    shorttitle = {Selecting the {Number} of {Clusters} $K$ with a {Stability} {Trade}-off},\n    doi = {10.48550/arXiv.2006.08530},\n    abstract = {Model selection is a major challenge in non-parametric clustering. There is no universally admitted way to evaluate clustering results for the obvious reason that no ground truth is available. The difficulty to find a universal evaluation criterion is a consequence of the ill-defined objective of clustering. In this perspective, clustering stability has emerged as a natural and model-agnostic principle: an algorithm should find stable structures in the data. If data sets are repeatedly sampled from the same underlying distribution, an algorithm should find similar partitions. However, stability alone is not well-suited to determine the number of clusters. For instance, it is unable to detect if the number of clusters is too small. We propose a new principle: a good clustering should be stable, and within each cluster, there should exist no stable partition. This principle leads to a novel clustering validation criterion based on between-cluster and within-cluster stability, overcoming limitations of previous stability-based methods. We empirically demonstrate the effectiveness of our criterion to select the number of clusters and compare it with existing methods. Code is available at https://github.com/FlorentF9/skstab.},\n    booktitle = {{PAKDD}},\n    author = {Mourer, Alex and Forest, Florent and Lebbah, Mustapha and Azzag, Hanane and Lacaille, Jérôme},\n    month = may,\n    year = {2023},\n    note = {arXiv:2006.08530 [cs, stat]},\n    keywords = {validity index,clustering,model selection,stability analysis},\n    url_Link = {http://arxiv.org/abs/2006.08530},\n    url_Paper = {http://arxiv.org/pdf/2006.08530.pdf},\n    bibbase_note = {<img src="assets/img/papers/stadion.png">}\n}\n\n
\n
\n\n\n
\n Model selection is a major challenge in non-parametric clustering. There is no universally admitted way to evaluate clustering results for the obvious reason that no ground truth is available. The difficulty to find a universal evaluation criterion is a consequence of the ill-defined objective of clustering. In this perspective, clustering stability has emerged as a natural and model-agnostic principle: an algorithm should find stable structures in the data. If data sets are repeatedly sampled from the same underlying distribution, an algorithm should find similar partitions. However, stability alone is not well-suited to determine the number of clusters. For instance, it is unable to detect if the number of clusters is too small. We propose a new principle: a good clustering should be stable, and within each cluster, there should exist no stable partition. This principle leads to a novel clustering validation criterion based on between-cluster and within-cluster stability, overcoming limitations of previous stability-based methods. We empirically demonstrate the effectiveness of our criterion to select the number of clusters and compare it with existing methods. Code is available at https://github.com/FlorentF9/skstab.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2021\n \n \n (2)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Deep Embedded Self-Organizing Map for Joint Representation Learning and Topology-Preserving Clustering.\n \n \n \n \n\n\n \n Florent Forest, Mustapha Lebbah, Hanene Azzag, & Jérôme Lacaille.\n\n\n \n\n\n\n Neural Computing and Applications. 2021.\n \n\n\n\n
\n\n\n\n \n \n \"Deep link\n  \n \n \n \"Deep paper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 104 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@article{forest2021deepembedded,\nabstract = {A recent research area in unsupervised learning is the combination of representation learning with deep neural networks and data clustering. The success of deep learning for supervised tasks is widely established. However, recent research has demonstrated how neural networks are able to learn representations to improve clustering in their intermediate feature space, using specific regularizations. By considering representation learning and clustering as a joint task, models learn clustering-friendly spaces and outperform two-stage approaches where dimensionality reduction and clustering are performed separately. Recently, this idea has been extended to topology-preserving clustering models, known as self-organizing maps (SOM). This work is a thorough study on the deep embedded self-organizing map (DESOM), a model composed of an autoencoder and a SOM layer, training jointly the code vectors and network weights to learn SOM-friendly representations. In other words, SOM induces a form a regularization to improve the quality of quantization and topology in latent space. After detailing the architecture, loss and training algorithm, we study hyperparameters with a series of experiments. Different SOM-based models are evaluated in terms of clustering, visualization and classification on benchmark datasets. We study benefits and trade-offs of joint representation learning and self-organization. DESOM achieves competitive results, requires no pretraining and produces topologically organized visualizations.},\nauthor = {Forest, Florent and Lebbah, Mustapha and Azzag, Hanene and Lacaille, J{\\'{e}}r{\\^{o}}me},\ndoi = {10.1007/s00521-021-06331-w},\nisbn = {0052102106},\njournal = {Neural Computing and Applications},\nkeywords = {autoencoder,clustering,deep learning,representation learning,self-organizing maps,visualization},\ntitle = {{Deep Embedded Self-Organizing Map for Joint Representation Learning and Topology-Preserving Clustering}},\nyear = {2021},\nurl_Link = {https://link.springer.com/article/10.1007/s00521-021-06331-w},\nurl_Paper = {https://www.researchgate.net/journal/Neural-Computing-and-Applications-1433-3058/publication/353679111_Deep_embedded_self-organizing_maps_for_joint_representation_learning_and_topology-preserving_clustering/links/610a2059169a1a0103daf991/Deep-embedded-self-organizing-maps-for-joint-representation-learning-and-topology-preserving-clustering.pdf},\nbibbase_note = {<img src="assets/img/papers/desom.png">}\n}\n\n
\n
\n\n\n
\n A recent research area in unsupervised learning is the combination of representation learning with deep neural networks and data clustering. The success of deep learning for supervised tasks is widely established. However, recent research has demonstrated how neural networks are able to learn representations to improve clustering in their intermediate feature space, using specific regularizations. By considering representation learning and clustering as a joint task, models learn clustering-friendly spaces and outperform two-stage approaches where dimensionality reduction and clustering are performed separately. Recently, this idea has been extended to topology-preserving clustering models, known as self-organizing maps (SOM). This work is a thorough study on the deep embedded self-organizing map (DESOM), a model composed of an autoencoder and a SOM layer, training jointly the code vectors and network weights to learn SOM-friendly representations. In other words, SOM induces a form a regularization to improve the quality of quantization and topology in latent space. After detailing the architecture, loss and training algorithm, we study hyperparameters with a series of experiments. Different SOM-based models are evaluated in terms of clustering, visualization and classification on benchmark datasets. We study benefits and trade-offs of joint representation learning and self-organization. DESOM achieves competitive results, requires no pretraining and produces topologically organized visualizations.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Unsupervised Learning of Data Representations and Cluster Structures: Applications to Large-scale Health Monitoring of Turbofan Aircraft Engines.\n \n \n \n \n\n\n \n Florent Forest.\n\n\n \n\n\n\n Ph.D. Thesis, Université Sorbonne Paris Nord, 2021.\n \n\n\n\n
\n\n\n\n \n \n \"Unsupervised link\n  \n \n \n \"Unsupervised paper\n  \n \n \n \"Unsupervised slides\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n  \n \n 36 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@phdthesis{forest2021unsupervised,\nauthor = {Forest, Florent},\npages = {326},\nschool = {Universit{\\'{e}} Sorbonne Paris Nord},\ntitle = {{Unsupervised Learning of Data Representations and Cluster Structures: Applications to Large-scale Health Monitoring of Turbofan Aircraft Engines}},\ntype = {PhD thesis},\nyear = {2021},\nurl_Link = {http://theses.fr/s194400},\nurl_Paper = {Forest2021-manuscrit.pdf},\nurl_Slides = {Forest2021-defense.pdf},\nbibbase_note = {<img src="assets/img/papers/these.png">}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2020\n \n \n (4)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n An Invariance-guided Stability Criterion for Time Series Clustering Validation.\n \n \n \n \n\n\n \n Florent Forest, Alex Mourer, Mustapha Lebbah, Hanane Azzag, & Jérôme Lacaille.\n\n\n \n\n\n\n In International Conference on Pattern Recognition (ICPR), 2020. \n \n\n\n\n
\n\n\n\n \n \n \"An link\n  \n \n \n \"An paper\n  \n \n \n \"An slides\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 25 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{forest2020invariance,\nabstract = {Time series clustering is a challenging task due to the specificities of this type of data. Temporal correlation and invariance to transformations such as shifting, warping or noise prevent the use of standard data mining methods. Time series clustering has been mostly studied under the angle of finding efficient algorithms and distance metrics adapted to the specific nature of time series data. Much less attention has been devoted to the general problem of model selection. Clustering stability has emerged as a universal and model-agnostic principle for clustering model selection. This principle can be stated as follows: an algorithm should find a structure in the data that is resilient to perturbation by sampling or noise. We propose to apply stability analysis to time series by leveraging prior knowledge on the nature and invariances of the data. These invariances determine the perturbation process used to assess stability. Based on a recently introduced criterion combining between-cluster and within-cluster stability, we propose an invariance-guided method for model selection, applicable to a wide range of clustering algorithms. Experiments conducted on artificial and benchmark data sets demonstrate the ability of our criterion to discover structure and select the correct number of clusters, whenever data invariances are known beforehand.},\nauthor = {Forest, Florent and Mourer, Alex and Lebbah, Mustapha and Azzag, Hanane and Lacaille, J{\\'{e}}r{\\^{o}}me},\nbooktitle = {International Conference on Pattern Recognition (ICPR)},\ntitle = {{An Invariance-guided Stability Criterion for Time Series Clustering Validation}},\nyear = {2020},\nurl_Link = {https://ieeexplore.ieee.org/abstract/document/9412020},\nurl_Paper = {ICPR-2020-InvarianceGuidedStabilityTSC-full-paper.pdf},\nurl_Slides = {pres-ICPR-2020.pdf},\nbibbase_note = {<img src="assets/img/papers/ts-stab.png">}\n}\n\n
\n
\n\n\n
\n Time series clustering is a challenging task due to the specificities of this type of data. Temporal correlation and invariance to transformations such as shifting, warping or noise prevent the use of standard data mining methods. Time series clustering has been mostly studied under the angle of finding efficient algorithms and distance metrics adapted to the specific nature of time series data. Much less attention has been devoted to the general problem of model selection. Clustering stability has emerged as a universal and model-agnostic principle for clustering model selection. This principle can be stated as follows: an algorithm should find a structure in the data that is resilient to perturbation by sampling or noise. We propose to apply stability analysis to time series by leveraging prior knowledge on the nature and invariances of the data. These invariances determine the perturbation process used to assess stability. Based on a recently introduced criterion combining between-cluster and within-cluster stability, we propose an invariance-guided method for model selection, applicable to a wide range of clustering algorithms. Experiments conducted on artificial and benchmark data sets demonstrate the ability of our criterion to discover structure and select the correct number of clusters, whenever data invariances are known beforehand.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Large-scale Vibration Monitoring of Aircraft Engines from Operational Data using Self-organized Models.\n \n \n \n \n\n\n \n Florent Forest, Quentin Cochard, Cecile Noyer, Adrien Cabut, Marc Joncour, Jérôme Lacaille, Mustapha Lebbah, & Hanene Azzag.\n\n\n \n\n\n\n In Annual Conference of the PHM Society, 2020. \n \n\n\n\n
\n\n\n\n \n \n \"Large-scale link\n  \n \n \n \"Large-scale paper\n  \n \n \n \"Large-scale slides\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 54 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{forest2020largescale,\nabstract = {Vibration analysis is an important component of industrial equipment health monitoring. Aircraft engines in particular are complex rotating machines where vibrations, mainly caused by unbalance, misalignment, or damaged bearings, put engine parts under dynamic structural stress. Thus, monitoring the vibratory behavior of engines is essential to detect anomalies and trends, avoid faults and improve availability. Intrinsic properties of parts can be described by the evolution of vibration as function of rotation speed, called a vibration signature. This work presents a methodology for large-scale vibration monitoring on operating civil aircraft engines, based on unsupervised learning algorithms and a flight recorder database. Firstly, we present a pipeline for massive extraction of vibration signatures from raw flight data, consisting in time-domain medium-frequency sensor measurements. Then, signatures are classified and visualized using interpretable self-organized clustering algorithms, yielding a visual cartography of vibration profiles. Domain experts can then extract various insights from resulting models. An abnormal temporal evolution of a signature gives early warning before failure of an engine. In a post-finding situation after an event has occurred, similar at-risk engines are detectable. The approach is global, end-to-end and scalable, which is yet uncommon in our industry, and has been tested on real flight data.},\nauthor = {Forest, Florent and Cochard, Quentin and Noyer, Cecile and Cabut, Adrien and Joncour, Marc and Lacaille, J{\\'{e}}r{\\^{o}}me and Lebbah, Mustapha and Azzag, Hanene},\nbooktitle = {Annual Conference of the PHM Society},\nkeywords = {aircraft engine, vibration analysis, health monitoring, big data, clustering, self-organizing map},\ndoi = {10.36001/phmconf.2020.v12i1.1131},\ntitle = {{Large-scale Vibration Monitoring of Aircraft Engines from Operational Data using Self-organized Models}},\nyear = {2020},\nurl_Link = {https://www.phmpapers.org/index.php/phmconf/article/view/1131},\nurl_Paper = {https://www.phmpapers.org/index.php/phmconf/article/download/1131/913},\nurl_Slides = {pres-PHM-2020.pdf},\nbibbase_note = {<img src="assets/img/papers/vib.png">}\n}\n\n
\n
\n\n\n
\n Vibration analysis is an important component of industrial equipment health monitoring. Aircraft engines in particular are complex rotating machines where vibrations, mainly caused by unbalance, misalignment, or damaged bearings, put engine parts under dynamic structural stress. Thus, monitoring the vibratory behavior of engines is essential to detect anomalies and trends, avoid faults and improve availability. Intrinsic properties of parts can be described by the evolution of vibration as function of rotation speed, called a vibration signature. This work presents a methodology for large-scale vibration monitoring on operating civil aircraft engines, based on unsupervised learning algorithms and a flight recorder database. Firstly, we present a pipeline for massive extraction of vibration signatures from raw flight data, consisting in time-domain medium-frequency sensor measurements. Then, signatures are classified and visualized using interpretable self-organized clustering algorithms, yielding a visual cartography of vibration profiles. Domain experts can then extract various insights from resulting models. An abnormal temporal evolution of a signature gives early warning before failure of an engine. In a post-finding situation after an event has occurred, similar at-risk engines are detectable. The approach is global, end-to-end and scalable, which is yet uncommon in our industry, and has been tested on real flight data.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n A Survey and Implementation of Performance Metrics for Self-Organized Maps.\n \n \n \n \n\n\n \n Florent Forest, Mustapha Lebbah, Hanane Azzag, & Jérôme Lacaille.\n\n\n \n\n\n\n November 2020.\n arXiv:2011.05847 [cs]\n\n\n\n
\n\n\n\n \n \n \"A link\n  \n \n \n \"A paper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 3 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@unpublished{forest2020survey,\nabstract = {Self-Organizing Map algorithms have been used for almost 40 years across various application domains such as biology, geology, healthcare, industry and humanities as an interpretable tool to explore, cluster and visualize high-dimensional data sets. In every application, practitioners need to know whether they can \\textit{trust} the resulting mapping, and perform model selection to tune algorithm parameters (e.g. the map size). Quantitative evaluation of self-organizing maps (SOM) is a subset of clustering validation, which is a challenging problem as such. Clustering model selection is typically achieved by using clustering validity indices. While they also apply to self-organized clustering models, they ignore the topology of the map, only answering the question: do the SOM code vectors approximate well the data distribution? Evaluating SOM models brings in the additional challenge of assessing their topology: does the mapping preserve neighborhood relationships between the map and the original data? The problem of assessing the performance of SOM models has already been tackled quite thoroughly in literature, giving birth to a family of quality indices incorporating neighborhood constraints, called \\textit{topographic} indices. Commonly used examples of such metrics are the topographic error, neighborhood preservation or the topographic product. However, open-source implementations are almost impossible to find. This is the issue we try to solve in this work: after a survey of existing SOM performance metrics, we implemented them in Python and widely used numerical libraries, and provide them as an open-source library, SOMperf. This paper introduces each metric available in our module along with usage examples.},\narchivePrefix = {arXiv},\narxivId = {arXiv:2011.05847},\nauthor = {Forest, Florent and Lebbah, Mustapha and Azzag, Hanane and Lacaille, J{\\'{e}}r{\\^{o}}me},\neprint = {arXiv:2011.05847},\nnote = {arXiv:2011.05847 [cs]},\ntitle = {{A Survey and Implementation of Performance Metrics for Self-Organized Maps}},\nyear = {2020},\nmonth = nov,\nurl_Link = {https://arxiv.org/abs/2011.05847},\nurl_Paper = {https://arxiv.org/pdf/2011.05847.pdf},\nbibbase_note = {<img src="assets/img/papers/somperf.png">}\n}\n\n
\n
\n\n\n
\n Self-Organizing Map algorithms have been used for almost 40 years across various application domains such as biology, geology, healthcare, industry and humanities as an interpretable tool to explore, cluster and visualize high-dimensional data sets. In every application, practitioners need to know whether they can trust the resulting mapping, and perform model selection to tune algorithm parameters (e.g. the map size). Quantitative evaluation of self-organizing maps (SOM) is a subset of clustering validation, which is a challenging problem as such. Clustering model selection is typically achieved by using clustering validity indices. While they also apply to self-organized clustering models, they ignore the topology of the map, only answering the question: do the SOM code vectors approximate well the data distribution? Evaluating SOM models brings in the additional challenge of assessing their topology: does the mapping preserve neighborhood relationships between the map and the original data? The problem of assessing the performance of SOM models has already been tackled quite thoroughly in literature, giving birth to a family of quality indices incorporating neighborhood constraints, called topographic indices. Commonly used examples of such metrics are the topographic error, neighborhood preservation or the topographic product. However, open-source implementations are almost impossible to find. This is the issue we try to solve in this work: after a survey of existing SOM performance metrics, we implemented them in Python and widely used numerical libraries, and provide them as an open-source library, SOMperf. This paper introduces each metric available in our module along with usage examples.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Carte SOM profonde : Apprentissage joint de représentations et auto-organisation.\n \n \n \n \n\n\n \n Florent Forest, Mustapha Lebbah, Hanene Azzag, & Jérôme Lacaille.\n\n\n \n\n\n\n In CAp2020: Conférence d'Apprentissage, 2020. \n \n\n\n\n
\n\n\n\n \n \n \"Carte link\n  \n \n \n \"Carte paper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 23 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{forest2020carte,\nabstract = {Dans la lign{\\'{e}}e des r{\\'{e}}centes avanc{\\'{e}}es en apprentissage profond de repr{\\'{e}}sentations pour le clustering, ce travail (pr{\\'{e}}c{\\'{e}}demment publi{\\'{e}} en anglais) pr{\\'{e}}sente le mod{\\`{e}}le DESOM (Deep Embedded SOM), combinant l'apprentisssage non supervis{\\'{e}} de repr{\\'{e}}sentations et d'une carte auto-organis{\\'{e}}e de Kohonen (SOM). Le mod{\\`{e}}le, compos{\\'{e}} d'un auto-encodeur et d'une couche SOM, est optimis{\\'{e}} conjointement, an de r{\\'{e}}gulariser l'espace latent et am{\\'{e}}liorer la performance de la carte SOM. Nous {\\'{e}}valuons les performances de classification et de visualisation ainsi que les b{\\'{e}}n{\\'{e}}fices de l'apprentissage joint. Mots-clef : carte auto-organis{\\'{e}}e, clustering, apprentissage profond, auto-encodeur.},\nauthor = {Forest, Florent and Lebbah, Mustapha and Azzag, Hanene and Lacaille, J{\\'{e}}r{\\^{o}}me},\nbooktitle = {CAp2020: Conf{\\'{e}}rence d'Apprentissage},\nkeywords = {autoencoder,clustering,deep learning,self-organizing map},\ntitle = {{Carte SOM profonde : Apprentissage joint de repr{\\'{e}}sentations et auto-organisation}},\nurl_Link = {https://hal.archives-ouvertes.fr/hal-02859997},\nurl_Paper = {https://hal.archives-ouvertes.fr/hal-02859997/document},\nyear = {2020}\n}\n\n
\n
\n\n\n
\n Dans la lignée des récentes avancées en apprentissage profond de représentations pour le clustering, ce travail (précédemment publié en anglais) présente le modèle DESOM (Deep Embedded SOM), combinant l'apprentisssage non supervisé de représentations et d'une carte auto-organisée de Kohonen (SOM). Le modèle, composé d'un auto-encodeur et d'une couche SOM, est optimisé conjointement, an de régulariser l'espace latent et améliorer la performance de la carte SOM. Nous évaluons les performances de classification et de visualisation ainsi que les bénéfices de l'apprentissage joint. Mots-clef : carte auto-organisée, clustering, apprentissage profond, auto-encodeur.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2019\n \n \n (2)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Deep Architectures for Joint Clustering and Visualization with Self-Organizing Maps.\n \n \n \n \n\n\n \n Florent Forest, Mustapha Lebbah, Hanane Azzag, & Jérôme Lacaille.\n\n\n \n\n\n\n In Workshop on Learning Data Representations for Clustering (LDRC), PAKDD, 2019. \n \n\n\n\n
\n\n\n\n \n \n \"Deep link\n  \n \n \n \"Deep paper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 38 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{forest2019deeparchitectures,\nabstract = {Recent research has demonstrated how deep neural networks are able to learn representations to improve data clustering. By considering representation learning and clustering as a joint task, models learn clustering-friendly spaces and achieve superior performance, com- pared with standard two-stage approaches where dimensionality reduc- tion and clustering are performed separately. We extend this idea to topology-preserving clustering models, known as self-organizing maps (SOM). First, we present the Deep Embedded Self-Organizing Map (DE- SOM), a model composed of a fully-connected autoencoder and a custom SOM layer, where the SOM code vectors are learnt jointly with the au- toencoder weights. Then, we show that this generic architecture can be extended to image and sequence data by using convolutional and recur- rent architectures, and present variants of these models. First results demonstrate advantages of the DESOM architecture in terms of cluster- ing performance, visualization and training time.},\nauthor = {Forest, Florent and Lebbah, Mustapha and Azzag, Hanane and Lacaille, J{\\'{e}}r{\\^{o}}me},\nbooktitle = {Workshop on Learning Data Representations for Clustering (LDRC), PAKDD},\ndoi = {10.1007/978-3-030-26142-9_10},\nkeywords = {autoencoder,clustering,deep learning,representation learning,self-organizing map},\ntitle = {{Deep Architectures for Joint Clustering and Visualization with Self-Organizing Maps}},\nyear = {2019},\nurl_Link = {https://link.springer.com/chapter/10.1007/978-3-030-26142-9_10},\nurl_Paper = {LDRC-2019-DeepArchitecturesJointClusteringVisualization-full-paper.pdf},\nbibbase_note = {<img src="assets/img/papers/convdesom.png">}\n}\n\n
\n
\n\n\n
\n Recent research has demonstrated how deep neural networks are able to learn representations to improve data clustering. By considering representation learning and clustering as a joint task, models learn clustering-friendly spaces and achieve superior performance, com- pared with standard two-stage approaches where dimensionality reduc- tion and clustering are performed separately. We extend this idea to topology-preserving clustering models, known as self-organizing maps (SOM). First, we present the Deep Embedded Self-Organizing Map (DE- SOM), a model composed of a fully-connected autoencoder and a custom SOM layer, where the SOM code vectors are learnt jointly with the au- toencoder weights. Then, we show that this generic architecture can be extended to image and sequence data by using convolutional and recur- rent architectures, and present variants of these models. First results demonstrate advantages of the DESOM architecture in terms of cluster- ing performance, visualization and training time.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Deep Embedded SOM: Joint Representation Learning and Self-Organization.\n \n \n \n \n\n\n \n Florent Forest, Mustapha Lebbah, Hanane Azzag, & Jérôme Lacaille.\n\n\n \n\n\n\n In European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN), 2019. \n \n\n\n\n
\n\n\n\n \n \n \"Deep link\n  \n \n \n \"Deep paper\n  \n \n \n \"Deep slides\n  \n \n \n \"Deep code\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 103 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{forest2019deepembedded,\nabstract = {In the wake of recent advances in joint clustering and deep learning, we introduce the Deep Embedded Self-Organizing Map, a model that jointly learns representations and the code vectors of a self-organizing map. Our model is composed of an autoencoder and a custom SOM layer that are optimized in a joint training procedure, motivated by the idea that the SOM prior could help learning SOM-friendly representations. We eval- uate SOM-based models in terms of clustering quality and unsupervised clustering accuracy, and study the benefits of joint training.},\nauthor = {Forest, Florent and Lebbah, Mustapha and Azzag, Hanane and Lacaille, J{\\'{e}}r{\\^{o}}me},\nbooktitle = {European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN)},\nkeywords = {autoencoder,clustering,deep learning,representation learning,self-organizing map},\ntitle = {{Deep Embedded SOM: Joint Representation Learning and Self-Organization}},\nyear = {2019},\nurl_Link = {https://www.esann.org/proceedings/2019},\nurl_Paper = {https://www.elen.ucl.ac.be/Proceedings/esann/esannpdf/es2019-30.pdf},\nurl_Slides = {ESANN-2019-DeepEmbeddedSOM-pres.pdf},\nurl_Code = {https://github.com/FlorentF9/DESOM},\nbibbase_note = {<img src="assets/img/papers/desom-maps.png">}\n}\n\n
\n
\n\n\n
\n In the wake of recent advances in joint clustering and deep learning, we introduce the Deep Embedded Self-Organizing Map, a model that jointly learns representations and the code vectors of a self-organizing map. Our model is composed of an autoencoder and a custom SOM layer that are optimized in a joint training procedure, motivated by the idea that the SOM prior could help learning SOM-friendly representations. We eval- uate SOM-based models in terms of clustering quality and unsupervised clustering accuracy, and study the benefits of joint training.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2018\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n A Generic and Scalable Pipeline for Large-Scale Analytics of Continuous Aircraft Engine Data.\n \n \n \n \n\n\n \n Florent Forest, Jérôme Lacaille, Mustapha Lebbah, & Hanane Azzag.\n\n\n \n\n\n\n In IEEE International Conference on Big Data, 2018. \n \n\n\n\n
\n\n\n\n \n \n \"A link\n  \n \n \n \"A paper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 9 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{forest2018generic,\nabstract = {A major application of data analytics for aircraft engine manufacturers is engine health monitoring, which consists in improving availability and operation of engines by leveraging operational data and past events. Traditional tools can no longer handle the increasing volume and velocity of data collected on modern aircraft. We propose a generic and scalable pipeline for large-scale analytics of operational data from a recent type of aircraft engine, oriented towards health monitoring applications. Based on Hadoop and Spark, our approach enables domain experts to scale their algorithms and extract features from tens of thousands of flights stored on a cluster. All computations are performed using the Spark framework, however custom functions and algorithms can be integrated without knowledge of distributed programming. Unsupervised learning algorithms are integrated for clustering and dimensionality reduction of the flight features, in order to allow efficient visualization and interpretation through a dedicated web application. The use case guiding our work is a methodology for engine fleet monitoring with a self-organizing map. Finally, this pipeline is meant to be end-to-end, fully customizable and ready for use in an industrial setting.},\nauthor = {Forest, Florent and Lacaille, J{\\'{e}}r{\\^{o}}me and Lebbah, Mustapha and Azzag, Hanane},\nbooktitle = {IEEE International Conference on Big Data},\ndoi = {10.1109/BigData.2018.8622297},\nisbn = {9781538650356},\nkeywords = {big data,aircraft engine,aviation,generic,hadoop,health monitoring,scalable,spark},\ntitle = {{A Generic and Scalable Pipeline for Large-Scale Analytics of Continuous Aircraft Engine Data}},\nyear = {2018},\nurl_Link = {https://ieeexplore.ieee.org/document/8622297},\nurl_Paper = {IEEEBigData-2018-ForestLacailleLebbahAzzag-full-paper.pdf},\nbibbase_note = {<img src="assets/img/papers/pipeline.png">}\n}\n\n
\n
\n\n\n
\n A major application of data analytics for aircraft engine manufacturers is engine health monitoring, which consists in improving availability and operation of engines by leveraging operational data and past events. Traditional tools can no longer handle the increasing volume and velocity of data collected on modern aircraft. We propose a generic and scalable pipeline for large-scale analytics of operational data from a recent type of aircraft engine, oriented towards health monitoring applications. Based on Hadoop and Spark, our approach enables domain experts to scale their algorithms and extract features from tens of thousands of flights stored on a cluster. All computations are performed using the Spark framework, however custom functions and algorithms can be integrated without knowledge of distributed programming. Unsupervised learning algorithms are integrated for clustering and dimensionality reduction of the flight features, in order to allow efficient visualization and interpretation through a dedicated web application. The use case guiding our work is a methodology for engine fleet monitoring with a self-organizing map. Finally, this pipeline is meant to be end-to-end, fully customizable and ready for use in an industrial setting.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2016\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n GammaLib and ctools: A software framework for the analysis of astronomical gamma-ray data.\n \n \n \n \n\n\n \n J. Knödlseder, M. Mayer, C. Deil, J. B. Cayrou, E. Owen, N. Kelley-Hoskins, C. Lu, R. Buehler, F. Forest, T. Louge, H. Siejkowski, K. Kosack, L. Gerard, A. Schulz, P. Martin, D. Sanchez, S. Ohm, T. Hassan, & S. Brau-Nogué.\n\n\n \n\n\n\n Astronomy and Astrophysics, 593: 1–19. 2016.\n \n\n\n\n
\n\n\n\n \n \n \"GammaLib link\n  \n \n \n \"GammaLib paper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 5 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n\n\n\n
\n
@article{knodlseder2016gammalib,\nabstract = {The field of gamma-ray astronomy has seen important progress during the last decade, yet to date no common software framework has been developed for the scientific analysis of gamma-ray telescope data. We propose to fill this gap by means of the GammaLib software, a generic library that we have developed to support the analysis of gamma-ray event data. GammaLib was written in C++ and all functionality is available in Python through an extension module. Based on this framework we have developed the ctools software package, a suite of software tools that enables flexible workflows to be built for the analysis of Imaging Air Cherenkov Telescope event data. The ctools are inspired by science analysis software available for existing high-energy astronomy instruments, and they follow the modular ftools model developed by the High Energy Astrophysics Science Archive Research Center. The ctools were written in Python and C++, and can be either used from the command line via shell scripts or directly from Python. In this paper we present the GammaLib and ctools software versions 1.0 that were released at the end of 2015. GammaLib and ctools are ready for the science analysis of Imaging Air Cherenkov Telescope event data, and also support the analysis of Fermi-LAT data and the exploitation of the COMPTEL legacy data archive. We propose using ctools as the science tools software for the Cherenkov Telescope Array Observatory.},\narchivePrefix = {arXiv},\narxivId = {1606.00393},\nauthor = {Kn{\\"{o}}dlseder, J. and Mayer, M. and Deil, C. and Cayrou, J. B. and Owen, E. and Kelley-Hoskins, N. and Lu, C. C. and Buehler, R. and Forest, F. and Louge, T. and Siejkowski, H. and Kosack, K. and Gerard, L. and Schulz, A. and Martin, P. and Sanchez, D. and Ohm, S. and Hassan, T. and Brau-Nogu{\\'{e}}, S.},\ndoi = {10.1051/0004-6361/201628822},\neprint = {1606.00393},\nissn = {14320746},\njournal = {Astronomy and Astrophysics},\nkeywords = {data analysis,virtual observatory tools},\npages = {1--19},\ntitle = {{GammaLib and ctools: A software framework for the analysis of astronomical gamma-ray data}},\nvolume = {593},\nyear = {2016},\nurl_Link = {https://www.aanda.org/articles/aa/abs/2016/09/aa28822-16/aa28822-16.html},\nurl_Paper = {https://www.aanda.org/articles/aa/pdf/2016/09/aa28822-16.pdf},\nbibbase_note = {<img src="assets/img/papers/gammalib.png">}\n}\n
\n
\n\n\n
\n The field of gamma-ray astronomy has seen important progress during the last decade, yet to date no common software framework has been developed for the scientific analysis of gamma-ray telescope data. We propose to fill this gap by means of the GammaLib software, a generic library that we have developed to support the analysis of gamma-ray event data. GammaLib was written in C++ and all functionality is available in Python through an extension module. Based on this framework we have developed the ctools software package, a suite of software tools that enables flexible workflows to be built for the analysis of Imaging Air Cherenkov Telescope event data. The ctools are inspired by science analysis software available for existing high-energy astronomy instruments, and they follow the modular ftools model developed by the High Energy Astrophysics Science Archive Research Center. The ctools were written in Python and C++, and can be either used from the command line via shell scripts or directly from Python. In this paper we present the GammaLib and ctools software versions 1.0 that were released at the end of 2015. GammaLib and ctools are ready for the science analysis of Imaging Air Cherenkov Telescope event data, and also support the analysis of Fermi-LAT data and the exploitation of the COMPTEL legacy data archive. We propose using ctools as the science tools software for the Cherenkov Telescope Array Observatory.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n\n\n\n
\n\n\n \n\n \n \n \n \n\n
\n"}; document.write(bibbase_data.data);