var bibbase_data = {"data":"\"Loading..\"\n\n
\n\n \n\n \n\n \n \n\n \n\n \n \n\n \n\n \n
\n generated by\n \n \"bibbase.org\"\n\n \n
\n \n\n
\n\n \n\n\n
\n\n Excellent! Next you can\n create a new website with this list, or\n embed it in an existing web page by copying & pasting\n any of the following snippets.\n\n
\n JavaScript\n (easiest)\n
\n \n <script src=\"https://bibbase.org/service/mendeley/aba9653c-d139-3f95-aad8-969c487ed2f3/group/aea1f410-b6a0-3014-afb8-bc5b4ad73524?jsonp=1&jsonp=1\"></script>\n \n
\n\n PHP\n
\n \n <?php\n $contents = file_get_contents(\"https://bibbase.org/service/mendeley/aba9653c-d139-3f95-aad8-969c487ed2f3/group/aea1f410-b6a0-3014-afb8-bc5b4ad73524?jsonp=1\");\n print_r($contents);\n ?>\n \n
\n\n iFrame\n (not recommended)\n
\n \n <iframe src=\"https://bibbase.org/service/mendeley/aba9653c-d139-3f95-aad8-969c487ed2f3/group/aea1f410-b6a0-3014-afb8-bc5b4ad73524?jsonp=1\"></iframe>\n \n
\n\n

\n For more details see the documention.\n

\n
\n
\n\n
\n\n This is a preview! To use this list on your own web site\n or create a new web site from it,\n create a free account. The file will be added\n and you will be able to edit it in the File Manager.\n We will show you instructions once you've created your account.\n
\n\n
\n\n

To the site owner:

\n\n

Action required! Mendeley is changing its\n API. In order to keep using Mendeley with BibBase past April\n 14th, you need to:\n

    \n
  1. renew the authorization for BibBase on Mendeley, and
  2. \n
  3. update the BibBase URL\n in your page the same way you did when you initially set up\n this page.\n
  4. \n
\n

\n\n

\n \n \n Fix it now\n

\n
\n\n
\n\n\n
\n \n \n
\n
\n  \n 2025\n \n \n (2)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Automatic Identification of Forest Areas in the ``Carolina'' Park Using ResNet50, EfficientNetB0 and VGG16: A Case Study.\n \n \n \n \n\n\n \n Guapaz, J.; Jervis, J., P.; Haro, D.; Padilla, J.; Guachi, R.; Peluffo-Ordóñez, D., H.; and Guachi-Guachi, L.\n\n\n \n\n\n\n In Florez, H.; and Astudillo, H., editor(s), Applied Informatics, pages 31-42, 2025. Springer Nature Switzerland\n \n\n\n\n
\n\n\n\n \n \n \"AutomaticWebsite\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{\n title = {Automatic Identification of Forest Areas in the ``Carolina'' Park Using ResNet50, EfficientNetB0 and VGG16: A Case Study},\n type = {inproceedings},\n year = {2025},\n pages = {31-42},\n websites = {https://link.springer.com/chapter/10.1007/978-3-031-75144-8_3},\n publisher = {Springer Nature Switzerland},\n city = {Cham},\n id = {0feb5a1a-f84d-3920-9540-8f59bb2f4515},\n created = {2024-11-12T17:54:48.400Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2024-11-12T17:54:48.400Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {10.1007/978-3-031-75144-8_3},\n source_type = {inproceedings},\n private_publication = {false},\n abstract = {This study explores the challenges of identifying forest areas in the ``Carolina'' Park in Quito, Ecuador, using Convolutional Neural Networks (CNN) and aerial imagery to support sustainable urban expansion plans. A dataset was constructed using 32\\$\\$\\backslash,\\backslashtimes \\backslash,\\$\\$\\texttimes32\\$\\$\\backslash,\\backslashtimes \\backslash,\\$\\$\\texttimes3 pixel patches extracted from 230 aerial images obtained from several videos captured by drones over the park. Three CNN models (ResNet50, EfficientNetB0 and VGG16) were trained to differentiate between forested, non-forested and hybrid areas. The methodology involved manual classification of 2100 patches into these three categories. The results showed that ResNet50 performed the best overall, with an accuracy of 76.66\\% \\textpm 8\\%, followed closely by VGG16, while EfficientNetB0 showed inferior performance on this specific dataset. Qualitative analysis of predictions on test images confirmed the effective identification of forest areas. These findings suggest that ResNet50 may be a suitable model for this task, demonstrating a high ability to learn and recognize patterns in forested areas through patch-based analysis, even with relatively small datasets derived from aerial drone imagery.},\n bibtype = {inproceedings},\n author = {Guapaz, Julian and Jervis, Juan Pablo and Haro, Diego and Padilla, Jefferson and Guachi, Robinson and Peluffo-Ordóñez, D H and Guachi-Guachi, Lorena},\n editor = {Florez, Hector and Astudillo, Hernán},\n booktitle = {Applied Informatics}\n}
\n
\n\n\n
\n This study explores the challenges of identifying forest areas in the ``Carolina'' Park in Quito, Ecuador, using Convolutional Neural Networks (CNN) and aerial imagery to support sustainable urban expansion plans. A dataset was constructed using 32\\$\\$\\backslash,\\backslashtimes \\backslash,\\$\\$\\texttimes32\\$\\$\\backslash,\\backslashtimes \\backslash,\\$\\$\\texttimes3 pixel patches extracted from 230 aerial images obtained from several videos captured by drones over the park. Three CNN models (ResNet50, EfficientNetB0 and VGG16) were trained to differentiate between forested, non-forested and hybrid areas. The methodology involved manual classification of 2100 patches into these three categories. The results showed that ResNet50 performed the best overall, with an accuracy of 76.66\\% \\textpm 8\\%, followed closely by VGG16, while EfficientNetB0 showed inferior performance on this specific dataset. Qualitative analysis of predictions on test images confirmed the effective identification of forest areas. These findings suggest that ResNet50 may be a suitable model for this task, demonstrating a high ability to learn and recognize patterns in forested areas through patch-based analysis, even with relatively small datasets derived from aerial drone imagery.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Automatic Differentiation Between Coriander and Parsley Using MobileNetV2.\n \n \n \n \n\n\n \n Páez, I.; Arévalo, J.; Martinez, M.; Molina, M.; Guachi, R.; Peluffo-Ordóñez, D., H.; and Guachi-Guachi, L.\n\n\n \n\n\n\n In Florez, H.; and Astudillo, H., editor(s), Applied Informatics, pages 18-30, 2025. Springer Nature Switzerland\n \n\n\n\n
\n\n\n\n \n \n \"AutomaticWebsite\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{\n title = {Automatic Differentiation Between Coriander and Parsley Using MobileNetV2},\n type = {inproceedings},\n year = {2025},\n pages = {18-30},\n websites = {https://link.springer.com/chapter/10.1007/978-3-031-75144-8_2},\n publisher = {Springer Nature Switzerland},\n city = {Cham},\n id = {295f4a04-2d1d-36e7-b290-da33944d17a2},\n created = {2024-11-12T18:03:35.797Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2024-11-12T18:03:35.797Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {10.1007/978-3-031-75144-8_2},\n source_type = {inproceedings},\n private_publication = {false},\n abstract = {Ecuadorian cuisine is very varied, with typical dishes from various regions often using fresh herbs. Among them, coriander and parsley are frequently used but often confused because of their similar appearance. This confusion can result in inexperienced or visually impaired people using the wrong herb, thus altering the flavor and unique qualities of the dishes. To address this problem, in this work, we present an image classification model capable of distinguishing between coriander and parsley leaves. Three architectures (MobileNetV2, EfficientNetB0 and InceptionV3) were evaluated to determine which offered the highest accuracy. MobileNetV2 proved to be the most efficient model, demonstrating superior stability and achieving greater than 99\\% accuracy.},\n bibtype = {inproceedings},\n author = {Páez, Ian and Arévalo, José and Martinez, Mateo and Molina, Martin and Guachi, Robinson and Peluffo-Ordóñez, D H and Guachi-Guachi, Lorena},\n editor = {Florez, Hector and Astudillo, Hernán},\n booktitle = {Applied Informatics}\n}
\n
\n\n\n
\n Ecuadorian cuisine is very varied, with typical dishes from various regions often using fresh herbs. Among them, coriander and parsley are frequently used but often confused because of their similar appearance. This confusion can result in inexperienced or visually impaired people using the wrong herb, thus altering the flavor and unique qualities of the dishes. To address this problem, in this work, we present an image classification model capable of distinguishing between coriander and parsley leaves. Three architectures (MobileNetV2, EfficientNetB0 and InceptionV3) were evaluated to determine which offered the highest accuracy. MobileNetV2 proved to be the most efficient model, demonstrating superior stability and achieving greater than 99\\% accuracy.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2024\n \n \n (3)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Instance Selection Framework for Alzheimer’s Disease Classification Using Multiple Regions of Interest and Atlas Integration.\n \n \n \n \n\n\n \n Castro-Silva., J., A.; Moreno-García., M.; Guachi-Guachi., L.; and Peluffo-Ordóñez., D., H.\n\n\n \n\n\n\n In Proceedings of the 13th International Conference on Pattern Recognition Applications and Methods - ICPRAM, pages 453-460, 2024. SciTePress\n \n\n\n\n
\n\n\n\n \n \n \"InstanceWebsite\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{\n title = {Instance Selection Framework for Alzheimer’s Disease Classification Using Multiple Regions of Interest and Atlas Integration},\n type = {inproceedings},\n year = {2024},\n pages = {453-460},\n websites = {https://www.scitepress.org/Link.aspx?doi=10.5220/0012469600003654},\n publisher = {SciTePress},\n institution = {INSTICC},\n id = {979ed068-0e34-3d58-a432-a1b8d2140266},\n created = {2024-03-06T14:15:39.886Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2024-03-06T14:15:39.886Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {icpram24},\n source_type = {conference},\n private_publication = {false},\n abstract = {Optimal selection of informative instances from a dataset is critical for constructing accurate predictive models. As databases expand, leveraging instance selection techniques becomes imperative to condense data into a more manageable size. This research unveils a novel framework designed to strategically identify and choose the most informative 2D brain image slices for Alzheimer’s disease classification. Such a framework integrates annotations from multiple regions of interest across multiple atlases. The proposed framework consists of six core components: 1) Atlas merging for ROI annotation and hemisphere separation. 2) Image preprocessing to extract informative slices. 3) Dataset construction to prevent data leakage, select subjects, and split data. 4) Data generation for memory-efficient batches. 5) Model construction for diverse classification training and testing. 6) Weighted ensemble for combining predictions from multiple models with a single learning algorithm. Our instanc e selection framework was applied to construct Transformer-based classification models, demonstrating an overall accuracy of approximately 98.33% in distinguishing between Cognitively Normal and Alzheimer’s cases at the subject level. It exhibited enhancements of 3.68%, 3.01%, 3.62% for sagittal, coronal, and axial planes respectively in comparison with the percentile technique.},\n bibtype = {inproceedings},\n author = {Castro-Silva., Juan A. and Moreno-García., Maria and Guachi-Guachi., Lorena and Peluffo-Ordóñez., Diego H.},\n doi = {10.5220/0012469600003654},\n booktitle = {Proceedings of the 13th International Conference on Pattern Recognition Applications and Methods - ICPRAM}\n}
\n
\n\n\n
\n Optimal selection of informative instances from a dataset is critical for constructing accurate predictive models. As databases expand, leveraging instance selection techniques becomes imperative to condense data into a more manageable size. This research unveils a novel framework designed to strategically identify and choose the most informative 2D brain image slices for Alzheimer’s disease classification. Such a framework integrates annotations from multiple regions of interest across multiple atlases. The proposed framework consists of six core components: 1) Atlas merging for ROI annotation and hemisphere separation. 2) Image preprocessing to extract informative slices. 3) Dataset construction to prevent data leakage, select subjects, and split data. 4) Data generation for memory-efficient batches. 5) Model construction for diverse classification training and testing. 6) Weighted ensemble for combining predictions from multiple models with a single learning algorithm. Our instanc e selection framework was applied to construct Transformer-based classification models, demonstrating an overall accuracy of approximately 98.33% in distinguishing between Cognitively Normal and Alzheimer’s cases at the subject level. It exhibited enhancements of 3.68%, 3.01%, 3.62% for sagittal, coronal, and axial planes respectively in comparison with the percentile technique.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Multiple Inputs and Mixed Data for Alzheimer’s Disease Classification Based on 3D Vision Transformer.\n \n \n \n \n\n\n \n Castro-Silva, J., A.; Moreno-García, M., N.; and Peluffo-Ordóñez, D., H.\n\n\n \n\n\n\n Mathematics, 12(17): 2720. 2024.\n \n\n\n\n
\n\n\n\n \n \n \"MultipleWebsite\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{\n title = {Multiple Inputs and Mixed Data for Alzheimer’s Disease Classification Based on 3D Vision Transformer},\n type = {article},\n year = {2024},\n pages = {2720},\n volume = {12},\n websites = {https://www.mdpi.com/2227-7390/12/17/2720},\n publisher = {MDPI},\n id = {67e1abb8-ae6d-3e39-bc38-f3a6afe5febf},\n created = {2024-09-04T16:45:13.004Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2024-09-04T16:45:13.004Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {castro2024multiple},\n source_type = {article},\n private_publication = {false},\n abstract = {The current methods for diagnosing Alzheimer’s Disease using Magnetic Resonance Imaging (MRI) have significant limitations. Many previous studies used 2D Transformers to analyze individual brain slices independently, potentially losing critical 3D contextual information. Region of interest-based models often focus on only a few brain regions despite Alzheimer’s affecting multiple areas. Additionally, most classification models rely on a single test, whereas diagnosing Alzheimer’s requires a multifaceted approach integrating diverse data sources for a more accurate assessment. This study introduces a novel methodology called the Multiple Inputs and Mixed Data 3D Vision Transformer (MIMD-3DVT). This method processes consecutive slices together to capture the feature dimensions and spatial information, fuses multiple 3D ROI imaging data inputs, and integrates mixed data from demographic factors, cognitive assessments, and brain imaging. The proposed methodology was experimentally evaluated using a combined dataset that included the Alzheimer’s Disease Neuroimaging Initiative (ADNI), the Australian Imaging, Biomarker, and Lifestyle Flagship Study of Ageing (AIBL), and the Open Access Series of Imaging Studies (OASIS). Our MIMD-3DVT, utilizing single or multiple ROIs, achieved an accuracy of 97.14%, outperforming the state-of-the-art methods in distinguishing between Normal Cognition and Alzheimer’s Disease.},\n bibtype = {article},\n author = {Castro-Silva, Juan A and Moreno-García, María N and Peluffo-Ordóñez, Diego H},\n journal = {Mathematics},\n number = {17}\n}
\n
\n\n\n
\n The current methods for diagnosing Alzheimer’s Disease using Magnetic Resonance Imaging (MRI) have significant limitations. Many previous studies used 2D Transformers to analyze individual brain slices independently, potentially losing critical 3D contextual information. Region of interest-based models often focus on only a few brain regions despite Alzheimer’s affecting multiple areas. Additionally, most classification models rely on a single test, whereas diagnosing Alzheimer’s requires a multifaceted approach integrating diverse data sources for a more accurate assessment. This study introduces a novel methodology called the Multiple Inputs and Mixed Data 3D Vision Transformer (MIMD-3DVT). This method processes consecutive slices together to capture the feature dimensions and spatial information, fuses multiple 3D ROI imaging data inputs, and integrates mixed data from demographic factors, cognitive assessments, and brain imaging. The proposed methodology was experimentally evaluated using a combined dataset that included the Alzheimer’s Disease Neuroimaging Initiative (ADNI), the Australian Imaging, Biomarker, and Lifestyle Flagship Study of Ageing (AIBL), and the Open Access Series of Imaging Studies (OASIS). Our MIMD-3DVT, utilizing single or multiple ROIs, achieved an accuracy of 97.14%, outperforming the state-of-the-art methods in distinguishing between Normal Cognition and Alzheimer’s Disease.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Novel hippocampus-centered methodology for informative instance selection in Alzheimer's disease data.\n \n \n \n\n\n \n Castro-Silva, J., A.; Moreno-García, M., N.; Guachi-Guachi, L.; and Peluffo-Ordóñez, D., H.\n\n\n \n\n\n\n Heliyon, 10(19): e37552. 2024.\n \n\n\n\n
\n\n\n\n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@article{\n title = {Novel hippocampus-centered methodology for informative instance selection in Alzheimer's disease data},\n type = {article},\n year = {2024},\n keywords = {Alzheimer's disease,Deep learning,Hippocampus,Instance selection},\n pages = {e37552},\n volume = {10},\n id = {b2d13133-eb4d-3fe0-bc3e-23479c4e0c95},\n created = {2024-09-25T22:33:25.171Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2024-09-25T22:33:25.171Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {CASTROSILVA2024e37552},\n source_type = {article},\n private_publication = {false},\n bibtype = {article},\n author = {Castro-Silva, Juan A and Moreno-García, María N and Guachi-Guachi, Lorena and Peluffo-Ordóñez, Diego H},\n doi = {https://doi.org/10.1016/j.heliyon.2024.e37552},\n journal = {Heliyon},\n number = {19}\n}
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2023\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Neural Networks on Noninvasive Electrocardiographic Imaging Reconstructions: Preliminary Results.\n \n \n \n \n\n\n \n Mayorca-Torres, D.; León-Salas, A., J.; and Peluffo-Ordoñez, D., H.\n\n\n \n\n\n\n In Botto-Tobar, M.; Gómez, O., S.; Rosero Miranda, R.; Díaz Cadena, A.; and Luna-Encalada, W., editor(s), Trends in Artificial Intelligence and Computer Engineering, pages 55-63, 2023. Springer Nature Switzerland\n \n\n\n\n
\n\n\n\n \n \n \"NeuralWebsite\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{\n title = {Neural Networks on Noninvasive Electrocardiographic Imaging Reconstructions: Preliminary Results},\n type = {inproceedings},\n year = {2023},\n pages = {55-63},\n websites = {https://link.springer.com/chapter/10.1007/978-3-031-25942-5_5},\n publisher = {Springer Nature Switzerland},\n city = {Cham},\n id = {e947ed90-061d-30f3-be14-0374b37b4f4f},\n created = {2023-02-13T23:39:20.520Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2023-02-13T23:39:20.520Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {10.1007/978-3-031-25942-5_5},\n source_type = {inproceedings},\n private_publication = {false},\n abstract = {In the reverse electrocardiography (ECG) problem, the objective is to reconstruct the heart's electrical activity from a set of body surface potentials by solving the direct model and the geometry of the torso. Over the years, researchers have used various approaches to solve this problem, from direct, iterative, probabilistic, and those based on deep learning. The interest of the latter, among the wide range of techniques, is because the complexity of the problem can be significantly reduced while increasing the precision of the estimation. In this article, we evaluate the performance of a deep learning-based neural network compared to the Tikhonov method of zero order (ZOT), first (FOT), and second (SOT). Preliminary results show an improvement in performance over real data when Pearson's correlation coefficient (CC) and (RMSE) are calculated. The CC's mean value and standard deviation for the proposed method were 0.960 (0.065), well above ZOT, which was 0.864 (0.047).},\n bibtype = {inproceedings},\n author = {Mayorca-Torres, Dagoberto and León-Salas, Alejandro José and Peluffo-Ordoñez, Diego Hernán},\n editor = {Botto-Tobar, Miguel and Gómez, Omar S and Rosero Miranda, Raul and Díaz Cadena, Angela and Luna-Encalada, Washington},\n booktitle = {Trends in Artificial Intelligence and Computer Engineering}\n}
\n
\n\n\n
\n In the reverse electrocardiography (ECG) problem, the objective is to reconstruct the heart's electrical activity from a set of body surface potentials by solving the direct model and the geometry of the torso. Over the years, researchers have used various approaches to solve this problem, from direct, iterative, probabilistic, and those based on deep learning. The interest of the latter, among the wide range of techniques, is because the complexity of the problem can be significantly reduced while increasing the precision of the estimation. In this article, we evaluate the performance of a deep learning-based neural network compared to the Tikhonov method of zero order (ZOT), first (FOT), and second (SOT). Preliminary results show an improvement in performance over real data when Pearson's correlation coefficient (CC) and (RMSE) are calculated. The CC's mean value and standard deviation for the proposed method were 0.960 (0.065), well above ZOT, which was 0.864 (0.047).\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2022\n \n \n (4)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Instance Selection on CNNs for Alzheimer’s Disease Classification from MRI.\n \n \n \n \n\n\n \n Castro-Silva., J.; Moreno-García., M.; Guachi-Guachi., L.; and Peluffo-Ordóñez., D.\n\n\n \n\n\n\n In Proceedings of the 11th International Conference on Pattern Recognition Applications and Methods - ICPRAM,, pages 330-337, 2022. SciTePress\n \n\n\n\n
\n\n\n\n \n \n \"InstanceWebsite\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{\n title = {Instance Selection on CNNs for Alzheimer’s Disease Classification from MRI},\n type = {inproceedings},\n year = {2022},\n pages = {330-337},\n websites = {https://www.scitepress.org/Link.aspx?doi=10.5220/0010900100003122},\n publisher = {SciTePress},\n institution = {INSTICC},\n id = {b1289f2e-9cba-3e3f-b692-0db0c4a8e9f2},\n created = {2022-02-22T01:00:12.425Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2022-02-22T01:00:12.425Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {icpram22},\n source_type = {conference},\n private_publication = {false},\n bibtype = {inproceedings},\n author = {Castro-Silva., J and Moreno-García., M and Guachi-Guachi., Lorena and Peluffo-Ordóñez., D},\n doi = {10.5220/0010900100003122},\n booktitle = {Proceedings of the 11th International Conference on Pattern Recognition Applications and Methods - ICPRAM,}\n}
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n ECG-Based Heartbeat Classification for Arrhythmia Detection Using Artificial Neural Networks.\n \n \n \n \n\n\n \n Cepeda, E.; Sánchez-Pozo, N., N.; Peluffo-Ordóñez, D., H.; González-Vergara, J.; and Almeida-Galárraga, D.\n\n\n \n\n\n\n In Gervasi, O.; Murgante, B.; Hendrix, E., M., T.; Taniar, D.; and Apduhan, B., O., editor(s), Computational Science and Its Applications -- ICCSA 2022, pages 247-259, 2022. Springer International Publishing\n \n\n\n\n
\n\n\n\n \n \n \"ECG-BasedWebsite\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 4 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{\n title = {ECG-Based Heartbeat Classification for Arrhythmia Detection Using Artificial Neural Networks},\n type = {inproceedings},\n year = {2022},\n pages = {247-259},\n websites = {https://link.springer.com/chapter/10.1007/978-3-031-10450-3_20},\n publisher = {Springer International Publishing},\n city = {Cham},\n id = {fe0bb748-7f2d-396d-be6a-7337af3ccdc3},\n created = {2022-07-16T03:47:10.517Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2022-07-16T03:47:10.517Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {10.1007/978-3-031-10450-3_20},\n source_type = {inproceedings},\n private_publication = {false},\n abstract = {Cardiovascular disease (CVD) has quickly grown in prevalence over the previous decade, becoming the major cause of human morbidity on a global scale. Due to the massive number of ECG data, manual analysis is regarded as a time-consuming, costly and prone to human error task. In the other hand, computational systems based on biomedical signal processing and machine learning techniques might be suited for supporting arrhythmia diagnostic processes, while solving some of those issues. In general, such systems involve five stages: acquisition, preprocessing, segmentation, characterization, and classification. Yet numerous fundamental aspects remain unresolved, including sensitivity to signal fluctuation, accuracy, computing cost, generalizability, and interpretability. In this context, the present study offers a comparative analysis of ECG signal classification using two artificial neural networks created by different machine learning frameworks. The neural nets were built into a pipeline that aims to strike an appropriate balance among signal robustness, variability, and accuracy. The proposed approach reaches up to 99\\% of overall accuracy for each register while keeping the computational cost low.},\n bibtype = {inproceedings},\n author = {Cepeda, Eduardo and Sánchez-Pozo, Nadia N and Peluffo-Ordóñez, Diego H and González-Vergara, Juan and Almeida-Galárraga, Diego},\n editor = {Gervasi, Osvaldo and Murgante, Beniamino and Hendrix, Eligius M T and Taniar, David and Apduhan, Bernady O},\n booktitle = {Computational Science and Its Applications -- ICCSA 2022}\n}
\n
\n\n\n
\n Cardiovascular disease (CVD) has quickly grown in prevalence over the previous decade, becoming the major cause of human morbidity on a global scale. Due to the massive number of ECG data, manual analysis is regarded as a time-consuming, costly and prone to human error task. In the other hand, computational systems based on biomedical signal processing and machine learning techniques might be suited for supporting arrhythmia diagnostic processes, while solving some of those issues. In general, such systems involve five stages: acquisition, preprocessing, segmentation, characterization, and classification. Yet numerous fundamental aspects remain unresolved, including sensitivity to signal fluctuation, accuracy, computing cost, generalizability, and interpretability. In this context, the present study offers a comparative analysis of ECG signal classification using two artificial neural networks created by different machine learning frameworks. The neural nets were built into a pipeline that aims to strike an appropriate balance among signal robustness, variability, and accuracy. The proposed approach reaches up to 99\\% of overall accuracy for each register while keeping the computational cost low.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n A Computer Vision Model to Identify the Incorrect Use of Face Masks for COVID-19 Awareness.\n \n \n \n \n\n\n \n Crespo, F.; Crespo, A.; Sierra-Martínez, L., M.; Peluffo-Ordóñez, D., H.; and Morocho-Cayamcela, M., E.\n\n\n \n\n\n\n Applied Sciences, 12(14). 2022.\n \n\n\n\n
\n\n\n\n \n \n \"AWebsite\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{\n title = {A Computer Vision Model to Identify the Incorrect Use of Face Masks for COVID-19 Awareness},\n type = {article},\n year = {2022},\n volume = {12},\n websites = {https://www.mdpi.com/2076-3417/12/14/6924},\n id = {70d7f700-db7a-39e4-8146-84cf0ff26274},\n created = {2022-07-24T01:29:08.758Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2022-07-24T01:29:08.758Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {app12146924},\n source_type = {article},\n private_publication = {false},\n abstract = {Face mask detection has become a great challenge in computer vision, demanding the coalition of technology with COVID-19 awareness. Researchers have proposed deep learning models to detect the use of face masks. However, the incorrect use of a face mask can be as harmful as not wearing any protection at all. In this paper, we propose a compound convolutional neural network (CNN) architecture based on two computer vision tasks: object localization to discover faces in images/videos, followed by an image classification CNN to categorize the faces and show if someone is using a face mask correctly, incorrectly, or not at all. The first CNN is built upon RetinaFace, a model to detect faces in images, whereas the second CNN uses a ResNet-18 architecture as a classification backbone. Our model enables an accurate identification of people who are not correctly following the COVID-19 healthcare recommendations on face mask use. To enable further global use of our technology, we have released both the dataset used to train the classification model and our proposed computer vision pipeline to the public, and optimized it for embedded systems deployment.},\n bibtype = {article},\n author = {Crespo, Fabricio and Crespo, Anthony and Sierra-Martínez, Luz Marina and Peluffo-Ordóñez, Diego Hernán and Morocho-Cayamcela, Manuel Eugenio},\n doi = {10.3390/app12146924},\n journal = {Applied Sciences},\n number = {14}\n}
\n
\n\n\n
\n Face mask detection has become a great challenge in computer vision, demanding the coalition of technology with COVID-19 awareness. Researchers have proposed deep learning models to detect the use of face masks. However, the incorrect use of a face mask can be as harmful as not wearing any protection at all. In this paper, we propose a compound convolutional neural network (CNN) architecture based on two computer vision tasks: object localization to discover faces in images/videos, followed by an image classification CNN to categorize the faces and show if someone is using a face mask correctly, incorrectly, or not at all. The first CNN is built upon RetinaFace, a model to detect faces in images, whereas the second CNN uses a ResNet-18 architecture as a classification backbone. Our model enables an accurate identification of people who are not correctly following the COVID-19 healthcare recommendations on face mask use. To enable further global use of our technology, we have released both the dataset used to train the classification model and our proposed computer vision pipeline to the public, and optimized it for embedded systems deployment.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Crop Classification Using Deep Learning: A Quick Comparative Study of Modern Approaches.\n \n \n \n \n\n\n \n Raki, H.; González-Vergara, J.; Aalaila, Y.; Elhamdi, M.; Bamansour, S.; Guachi-Guachi, L.; and Peluffo-Ordoñez, D., H.\n\n\n \n\n\n\n In Florez, H.; and Gomez, H., editor(s), Applied Informatics, pages 31-44, 2022. Springer International Publishing\n \n\n\n\n
\n\n\n\n \n \n \"CropWebsite\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{\n title = {Crop Classification Using Deep Learning: A Quick Comparative Study of Modern Approaches},\n type = {inproceedings},\n year = {2022},\n pages = {31-44},\n websites = {https://link.springer.com/chapter/10.1007/978-3-031-19647-8_3},\n publisher = {Springer International Publishing},\n city = {Cham},\n id = {d0ea8cbb-e8dd-3b80-8451-9b858e90db31},\n created = {2023-03-01T20:18:51.423Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2023-03-01T20:18:51.423Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {10.1007/978-3-031-19647-8_3},\n source_type = {inproceedings},\n private_publication = {false},\n abstract = {Automatic crop classification using new technologies is recognized as one of the most important assets in today's smart farming improvement. Investments in technology and innovation are key issues for shaping agricultural productivity as well as the inclusiveness and sustainability of the global agricultural transformation. Digital image processing (DIP) has been widely adopted in this field, by merging Unmanned Aerial Vehicle (UAV) based remote sensing and deep learning (DL) as a powerful tool for crop classification. Despite the wide range of alternatives, the proper selection of a DL approach is still an open and challenging issue. In this work, we carry out an exhaustive performance evaluation of three remarkable and lightweight DL approaches, namely: Visual Geometry Group (VGG), Residual Neural Network (ResNet) and Inception V3, tested on high resolution agriculture crop images dataset. Experimental results show that InceptionV3 outperforms VGG and ResNet in terms of precision (0,92), accuracy (0,97), recall (0,91), AUC (0,98), PCR (0,97), and F1 (0,91).},\n bibtype = {inproceedings},\n author = {Raki, Hind and González-Vergara, Juan and Aalaila, Yahya and Elhamdi, Mouad and Bamansour, Sami and Guachi-Guachi, Lorena and Peluffo-Ordoñez, Diego H},\n editor = {Florez, Hector and Gomez, Henry},\n booktitle = {Applied Informatics}\n}
\n
\n\n\n
\n Automatic crop classification using new technologies is recognized as one of the most important assets in today's smart farming improvement. Investments in technology and innovation are key issues for shaping agricultural productivity as well as the inclusiveness and sustainability of the global agricultural transformation. Digital image processing (DIP) has been widely adopted in this field, by merging Unmanned Aerial Vehicle (UAV) based remote sensing and deep learning (DL) as a powerful tool for crop classification. Despite the wide range of alternatives, the proper selection of a DL approach is still an open and challenging issue. In this work, we carry out an exhaustive performance evaluation of three remarkable and lightweight DL approaches, namely: Visual Geometry Group (VGG), Residual Neural Network (ResNet) and Inception V3, tested on high resolution agriculture crop images dataset. Experimental results show that InceptionV3 outperforms VGG and ResNet in terms of precision (0,92), accuracy (0,97), recall (0,91), AUC (0,98), PCR (0,97), and F1 (0,91).\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2021\n \n \n (2)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Enhanced Convolutional-Neural-Network Architecture for Crop Classification.\n \n \n \n \n\n\n \n Moreno-revelo, M., Y.; Guachi-guachi, L.; Gómez-mendoza, J., B.; Revelo-fuelagán, J.; and Peluffo-ordóñez, D., H.\n\n\n \n\n\n\n Applied Sciences,1-23. 2021.\n \n\n\n\n
\n\n\n\n \n \n \"EnhancedPaper\n  \n \n \n \"EnhancedWebsite\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{\n title = {Enhanced Convolutional-Neural-Network Architecture for Crop Classification},\n type = {article},\n year = {2021},\n pages = {1-23},\n websites = {https://www.mdpi.com/2076-3417/11/9/4292},\n id = {24a0c385-d592-3a29-a7e6-33a2d377a118},\n created = {2022-02-02T06:07:46.014Z},\n file_attached = {true},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2022-02-02T06:07:47.387Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n private_publication = {false},\n abstract = {Automatic crop identification and monitoring is a key element in enhancing food production processes as well as diminishing the related environmental impact. Although several efficient deep learning techniques have emerged in the field of multispectral imagery analysis, the crop classification problem still needs more accurate solutions. This work introduces a competitive methodology for crop classification from multispectral satellite imagery mainly using an enhanced 2D convolutional neural network (2D-CNN) designed at a smaller-scale architecture, as well as a novel post-processing step. The proposed methodology contains four steps: image stacking, patch extraction, classification model design (based on a 2D-CNN architecture), and post-processing. First, the images are stacked to increase the number of features. Second, the input images are split into patches and fed into the 2D-CNN model. Then, the 2D-CNN model is constructed within a small-scale framework, and properly trained to recognize 10 different types of crops. Finally, a post-processing step is performed in order to reduce the classification error caused by lower-spatial-resolution images. Experiments were carried over the so-named Campo Verde database, which consists of a set of satellite images captured by Landsat and Sentinel satellites from the municipality of Campo Verde, Brazil. In contrast to the maximum accuracy values reached by remarkable works reported in the literature (amounting to an overall accuracy of about 81%, a f1 score of 75.89%, and average accuracy of 73.35%), the proposed methodology achieves a competitive overall accuracy of 81.20%, a f1 score of 75.89%, and an average accuracy of 88.72% when classifying 10 different crops, while ensuring an adequate trade-off between the number of multiply-accumulate operations (MACs) and accuracy. Furthermore, given its ability to effectively classify patches from two image sequences, this methodology may result appealing for other real-world applications, such as the classification of urban materials.},\n bibtype = {article},\n author = {Moreno-revelo, Mónica Y and Guachi-guachi, Lorena and Gómez-mendoza, Juan Bernardo and Revelo-fuelagán, Javier and Peluffo-ordóñez, Diego H},\n journal = {Applied Sciences}\n}
\n
\n\n\n
\n Automatic crop identification and monitoring is a key element in enhancing food production processes as well as diminishing the related environmental impact. Although several efficient deep learning techniques have emerged in the field of multispectral imagery analysis, the crop classification problem still needs more accurate solutions. This work introduces a competitive methodology for crop classification from multispectral satellite imagery mainly using an enhanced 2D convolutional neural network (2D-CNN) designed at a smaller-scale architecture, as well as a novel post-processing step. The proposed methodology contains four steps: image stacking, patch extraction, classification model design (based on a 2D-CNN architecture), and post-processing. First, the images are stacked to increase the number of features. Second, the input images are split into patches and fed into the 2D-CNN model. Then, the 2D-CNN model is constructed within a small-scale framework, and properly trained to recognize 10 different types of crops. Finally, a post-processing step is performed in order to reduce the classification error caused by lower-spatial-resolution images. Experiments were carried over the so-named Campo Verde database, which consists of a set of satellite images captured by Landsat and Sentinel satellites from the municipality of Campo Verde, Brazil. In contrast to the maximum accuracy values reached by remarkable works reported in the literature (amounting to an overall accuracy of about 81%, a f1 score of 75.89%, and average accuracy of 73.35%), the proposed methodology achieves a competitive overall accuracy of 81.20%, a f1 score of 75.89%, and an average accuracy of 88.72% when classifying 10 different crops, while ensuring an adequate trade-off between the number of multiply-accumulate operations (MACs) and accuracy. Furthermore, given its ability to effectively classify patches from two image sequences, this methodology may result appealing for other real-world applications, such as the classification of urban materials.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Comparison of current deep convolutional neural networks for the segmentation of breast masses in mammograms.\n \n \n \n \n\n\n \n Anaya-Isaza, A.; Mera-Jiménez, L.; Cabrera-Chavarro, J.; Guachi-Guachi, L.; Peluffo-Ordóñez, D.; and Rios-Patiño, J.\n\n\n \n\n\n\n IEEE Access. 2021.\n \n\n\n\n
\n\n\n\n \n \n \"ComparisonWebsite\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{\n title = {Comparison of current deep convolutional neural networks for the segmentation of breast masses in mammograms},\n type = {article},\n year = {2021},\n websites = {https://ieeexplore.ieee.org/document/9614200},\n id = {c9a3316f-2543-320d-b8bb-5ed45ce86aab},\n created = {2022-02-02T06:07:46.249Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2022-02-02T06:07:46.249Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {9614200},\n source_type = {article},\n private_publication = {false},\n abstract = {Breast cancer causes approximately 684,996 deaths worldwide, making it the leading cause of female cancer mortality. However, these figures can be reduced with early diagnosis through mammographic imaging, allowing for the timely and effective treatment of this disease. To establish the best tools for contributing to the automatic diagnosis of breast cancer, different deep learning (DL) architectures were compared in terms of breast lesion segmentation, lesion type classification, and degree of suspicion of malignancy tests. The tasks were completed with state-of-the-art architectures and backbones. Initially, during segmentation, the base UNet, Visual Geometry Group 19 (VGG19), InceptionResNetV2, EfficientNet, MobileNetv2, ResNet, ResNeXt, MultiResUNet, linkNet-VGG19, DenseNet, SEResNet and SeResNeXt architectures were compared, where “Res” denotes a residual network. In addition, training was performed with 5 of the most advanced loss functions and validated by the Dice coefficient, sensitivity, and specificity. The proposed models achieved Dice values above 90%, with the EfficientNet architecture achieving 94.75% and 99% accuracy on the two tasks. Subsequently, classification was addressed with the ResNet50V2, VGG19, InceptionResNetV2, DenseNet121, InceptionV3, Xception and EfficientNetB7 networks. The proposed models achieved 96.97% and 97.73% accuracy through the VGG19 and ResNet50V2 networks on the lesion classification and degree of suspicion tasks, respectively. All three tasks were addressed with open-access databases, including the Digital Database for Screening Mammography (DDSM), the Mammographic Image Analysis Society (MIAS) database, and INbreast.},\n bibtype = {article},\n author = {Anaya-Isaza, Andrés and Mera-Jiménez, Leonel and Cabrera-Chavarro, Johan and Guachi-Guachi, Lorena and Peluffo-Ordóñez, Diego and Rios-Patiño, Jorge},\n doi = {10.1109/ACCESS.2021.3127862},\n journal = {IEEE Access}\n}
\n
\n\n\n
\n Breast cancer causes approximately 684,996 deaths worldwide, making it the leading cause of female cancer mortality. However, these figures can be reduced with early diagnosis through mammographic imaging, allowing for the timely and effective treatment of this disease. To establish the best tools for contributing to the automatic diagnosis of breast cancer, different deep learning (DL) architectures were compared in terms of breast lesion segmentation, lesion type classification, and degree of suspicion of malignancy tests. The tasks were completed with state-of-the-art architectures and backbones. Initially, during segmentation, the base UNet, Visual Geometry Group 19 (VGG19), InceptionResNetV2, EfficientNet, MobileNetv2, ResNet, ResNeXt, MultiResUNet, linkNet-VGG19, DenseNet, SEResNet and SeResNeXt architectures were compared, where “Res” denotes a residual network. In addition, training was performed with 5 of the most advanced loss functions and validated by the Dice coefficient, sensitivity, and specificity. The proposed models achieved Dice values above 90%, with the EfficientNet architecture achieving 94.75% and 99% accuracy on the two tasks. Subsequently, classification was addressed with the ResNet50V2, VGG19, InceptionResNetV2, DenseNet121, InceptionV3, Xception and EfficientNetB7 networks. The proposed models achieved 96.97% and 97.73% accuracy through the VGG19 and ResNet50V2 networks on the lesion classification and degree of suspicion tasks, respectively. All three tasks were addressed with open-access databases, including the Digital Database for Screening Mammography (DDSM), the Mammographic Image Analysis Society (MIAS) database, and INbreast.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2019\n \n \n (3)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Artificial Neural Networks for Bottled Water Demand Forecasting: A Small Business Case Study.\n \n \n \n \n\n\n \n Herrera-Granda, I., D.; Chicaiza-Ipiales, J., A.; Herrera-Granda, E., P.; Lorente-Leyva, L., L.; Caraguay-Procel, J., A.; García-Santillán, I., D.; and Peluffo-Ordóñez, D., H.\n\n\n \n\n\n\n Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), pages 362-373. 2019.\n \n\n\n\n
\n\n\n\n \n \n \"LectureWebsite\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inbook{\n type = {inbook},\n year = {2019},\n keywords = {Artificial neural networks,Long-term demand forecasting,Small business},\n pages = {362-373},\n websites = {http://link.springer.com/10.1007/978-3-030-20518-8_31},\n id = {d0ee1709-f877-3515-a99a-f05f24797555},\n created = {2022-02-02T06:07:46.500Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2022-02-02T06:07:46.500Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {Herrera-Granda2019a},\n private_publication = {false},\n abstract = {This paper shows a neural networks-based demand forecasting model designed for a small manufacturer of bottled water in Ecuador, which currently doesn’t have adequate demand forecast methodologies, causing problems of customer orders non-compliance, inventory excess and economic losses. However, by working with accurate predictions, the manufacturer will have an anticipated vision of future needs in order to satisfy the demand for manufactured products, in other words, to guarantee on time and reasonable use of the resources. To solve the problems that this small manufacturer has to face a historic demand data acquisition process was done through the last 36 months costumer order records. In the construction of the historical time series, that was analyzed, demand dates and volumes were established as input variables. Then the design of forecast models was done, based on classical methods and multi-layer neural networks, which were evaluated by means of quantitative error indicators. The application of these methods was done through the R programming language. After this, a stage of training and improvement of the network is included, it was evaluated against the results of the classic forecasting methods, and the next 12 months were predicted by means of the best obtained model. Finally, the feasibility of the use of neural networks in the forecast of demand for purified water bottles, is demonstrated.},\n bibtype = {inbook},\n author = {Herrera-Granda, Israel D. and Chicaiza-Ipiales, Joselyn A. and Herrera-Granda, Erick P. and Lorente-Leyva, Leandro L. and Caraguay-Procel, Jorge A. and García-Santillán, Iván D. and Peluffo-Ordóñez, Diego H.},\n doi = {10.1007/978-3-030-20518-8_31},\n chapter = {Artificial Neural Networks for Bottled Water Demand Forecasting: A Small Business Case Study},\n title = {Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)}\n}
\n
\n\n\n
\n This paper shows a neural networks-based demand forecasting model designed for a small manufacturer of bottled water in Ecuador, which currently doesn’t have adequate demand forecast methodologies, causing problems of customer orders non-compliance, inventory excess and economic losses. However, by working with accurate predictions, the manufacturer will have an anticipated vision of future needs in order to satisfy the demand for manufactured products, in other words, to guarantee on time and reasonable use of the resources. To solve the problems that this small manufacturer has to face a historic demand data acquisition process was done through the last 36 months costumer order records. In the construction of the historical time series, that was analyzed, demand dates and volumes were established as input variables. Then the design of forecast models was done, based on classical methods and multi-layer neural networks, which were evaluated by means of quantitative error indicators. The application of these methods was done through the R programming language. After this, a stage of training and improvement of the network is included, it was evaluated against the results of the classic forecasting methods, and the next 12 months were predicted by means of the best obtained model. Finally, the feasibility of the use of neural networks in the forecast of demand for purified water bottles, is demonstrated.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Artificial Neural Networks for Urban Water Demand Forecasting: A Case Study.\n \n \n \n \n\n\n \n Lorente-Leyva, L., L.; Pavón-Valencia, J., F.; Montero-Santos, Y.; Herrera-Granda, I., D.; Herrera-Granda, E., P.; and Peluffo-Ordóñez, D., H.\n\n\n \n\n\n\n Journal of Physics: Conference Series, 1284: 012004. 8 2019.\n \n\n\n\n
\n\n\n\n \n \n \"ArtificialWebsite\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{\n title = {Artificial Neural Networks for Urban Water Demand Forecasting: A Case Study},\n type = {article},\n year = {2019},\n pages = {012004},\n volume = {1284},\n websites = {https://iopscience.iop.org/article/10.1088/1742-6596/1284/1/012004},\n month = {8},\n id = {d25efbb9-d760-3246-ad23-47aedb6830b6},\n created = {2022-02-02T06:07:46.757Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2022-02-02T06:07:46.757Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {Lorente-Leyva2019a},\n private_publication = {false},\n abstract = {This paper presents an application of an artificial neural network model in forecasting urban water demand using MATLAB software. Considering that in any planning process, the demand forecast plays a fundamental role, being one of the premises to organize and control a set of activities or processes. The versatility of the short, medium and long-term prediction that is provided to the company that offers the water distribution service to determine the supply capacity, maintenance activities, and system improvements as a strategic planning tool. Shown to improve network performance by using time series water demand data, the model can provide excellent fit and forecast without relying on the explicit inclusion of climatic factors and number of consumers. The excellent accuracy of the model indicates the effectiveness of forecasting over different time horizons. Finally, the results obtained from the Artificial Neural Network are compared with traditional statistical models.},\n bibtype = {article},\n author = {Lorente-Leyva, Leandro L. and Pavón-Valencia, Jairo F. and Montero-Santos, Yakcleem and Herrera-Granda, Israel D and Herrera-Granda, Erick P. and Peluffo-Ordóñez, Diego H.},\n doi = {10.1088/1742-6596/1284/1/012004},\n journal = {Journal of Physics: Conference Series}\n}
\n
\n\n\n
\n This paper presents an application of an artificial neural network model in forecasting urban water demand using MATLAB software. Considering that in any planning process, the demand forecast plays a fundamental role, being one of the premises to organize and control a set of activities or processes. The versatility of the short, medium and long-term prediction that is provided to the company that offers the water distribution service to determine the supply capacity, maintenance activities, and system improvements as a strategic planning tool. Shown to improve network performance by using time series water demand data, the model can provide excellent fit and forecast without relying on the explicit inclusion of climatic factors and number of consumers. The excellent accuracy of the model indicates the effectiveness of forecasting over different time horizons. Finally, the results obtained from the Artificial Neural Network are compared with traditional statistical models.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Artificial Neural Networks in the Demand Forecasting of a Metal-Mechanical Industry.\n \n \n \n \n\n\n \n L. Lorente-Leyva, L.; R. Patino-Alarcon, D.; Montero-Santos, Y.; D. Herrera-Granda, I.; H. Peluffo-Ordonez, D.; M. Lastre-Aleaga, A.; and Cordoves-Garcia, A.\n\n\n \n\n\n\n Journal of Engineering and Applied Sciences, 15(1): 81-87. 10 2019.\n \n\n\n\n
\n\n\n\n \n \n \"ArtificialWebsite\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{\n title = {Artificial Neural Networks in the Demand Forecasting of a Metal-Mechanical Industry},\n type = {article},\n year = {2019},\n pages = {81-87},\n volume = {15},\n websites = {http://www.medwelljournals.com/abstract/?doi=jeasci.2020.81.87},\n month = {10},\n day = {25},\n id = {d2a33562-d740-37c2-a6df-cc39c178a563},\n created = {2022-02-02T06:07:46.991Z},\n file_attached = {false},\n profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},\n group_id = {aea1f410-b6a0-3014-afb8-bc5b4ad73524},\n last_modified = {2022-02-02T06:07:46.991Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {L.Lorente-Leyva2019},\n private_publication = {false},\n abstract = {This research presents an application of artificial neural networks in demand forecasting by using MATLAB Software. Keeping in mind that in any planning process forecasts play a fundamental role, being one of the bases for; planning, organizing and controlling production. It gives priority to the most critical nodes and their key activities, so that, the decisions made about them will generate the greatest possible positive impact. The methodology applied demonstrates the quality of the solutions found which are compared with traditional statistical methods to demonstrate the value of the solution proposed. When the results show that the minimum quadratic error is reached with the application of artificial neural networks, a better performance is obtained. Therefore, a suitable horizon is established for the planification and decision making in the metal-mechanical industry for the use of artificial intelligence in the production processes.},\n bibtype = {article},\n author = {L. Lorente-Leyva, Leandro and R. Patino-Alarcon, Delio and Montero-Santos, Yakcleem and D. Herrera-Granda, Israel and H. Peluffo-Ordonez, Diego and M. Lastre-Aleaga, Arlys and Cordoves-Garcia, Alexis},\n doi = {10.36478/jeasci.2020.81.87},\n journal = {Journal of Engineering and Applied Sciences},\n number = {1}\n}
\n
\n\n\n
\n This research presents an application of artificial neural networks in demand forecasting by using MATLAB Software. Keeping in mind that in any planning process forecasts play a fundamental role, being one of the bases for; planning, organizing and controlling production. It gives priority to the most critical nodes and their key activities, so that, the decisions made about them will generate the greatest possible positive impact. The methodology applied demonstrates the quality of the solutions found which are compared with traditional statistical methods to demonstrate the value of the solution proposed. When the results show that the minimum quadratic error is reached with the application of artificial neural networks, a better performance is obtained. Therefore, a suitable horizon is established for the planification and decision making in the metal-mechanical industry for the use of artificial intelligence in the production processes.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n\n\n\n
\n\n\n \n\n \n \n \n \n\n
\n"}; document.write(bibbase_data.data);