var bibbase_data = {"data":"\"Loading..\"\n\n
\n\n \n\n \n\n \n \n\n \n\n \n \n\n \n\n \n
\n generated by\n \n \"bibbase.org\"\n\n \n
\n \n\n
\n\n \n\n\n
\n\n Excellent! Next you can\n create a new website with this list, or\n embed it in an existing web page by copying & pasting\n any of the following snippets.\n\n
\n JavaScript\n (easiest)\n
\n \n <script src=\"https://bibbase.org/show?bib=http%3A%2F%2Flearn.inf.puc-rio.br%2Ffiles%2Fall.bib&jsonp=1&theme=side&jsonp=1\"></script>\n \n
\n\n PHP\n
\n \n <?php\n $contents = file_get_contents(\"https://bibbase.org/show?bib=http%3A%2F%2Flearn.inf.puc-rio.br%2Ffiles%2Fall.bib&jsonp=1&theme=side\");\n print_r($contents);\n ?>\n \n
\n\n iFrame\n (not recommended)\n
\n \n <iframe src=\"https://bibbase.org/show?bib=http%3A%2F%2Flearn.inf.puc-rio.br%2Ffiles%2Fall.bib&jsonp=1&theme=side\"></iframe>\n \n
\n\n

\n For more details see the documention.\n

\n
\n
\n\n
\n\n This is a preview! To use this list on your own web site\n or create a new web site from it,\n create a free account. The file will be added\n and you will be able to edit it in the File Manager.\n We will show you instructions once you've created your account.\n
\n\n
\n\n

To the site owner:

\n\n

Action required! Mendeley is changing its\n API. In order to keep using Mendeley with BibBase past April\n 14th, you need to:\n

    \n
  1. renew the authorization for BibBase on Mendeley, and
  2. \n
  3. update the BibBase URL\n in your page the same way you did when you initially set up\n this page.\n
  4. \n
\n

\n\n

\n \n \n Fix it now\n

\n
\n\n
\n\n\n
\n \n \n
\n
\n  \n 2019\n \n \n (8)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n A Deep Convolutional Network for Seismic Shot-Gather Image Quality Classification.\n \n \n \n\n\n \n Betine, E.; Busson, A.; Milidiú, R.; Colcher, S.; Dias, B.; and Bulcão, A.\n\n\n \n\n\n\n December 2019.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@misc{betine_deep_2019,\n  title = {A {Deep} {Convolutional} {Network} for {Seismic} {Shot}-{Gather} {Image} {Quality} {Classification}},\n  abstract = {Deep Learning-based models such as Convolutional Neural Networks, have led to significant advancements in several areas of computing applications. Seismogram quality assurance is a relevant Geophysics task, since in the early stages of seismic processing, we are required to identify and fix noisy sail lines. In this work, we introduce a real-world seismogram quality classification dataset based on 6,613 examples, manually labeled by human experts as good, bad or ugly, according to their noise intensity. This dataset is used to train a CNN classifier for seismic shot-gathers quality prediction. In our empirical evaluation, we observe an F1-score of 93.56\\% in the test set.},\n  author = {Betine, Eduardo and Busson, Antonio and Milidiú, Ruy and Colcher, Sérgio and Dias, Bruno and Bulcão, André},\n  month = dec,\n  year = {2019}\n}\n\n
\n
\n\n\n
\n Deep Learning-based models such as Convolutional Neural Networks, have led to significant advancements in several areas of computing applications. Seismogram quality assurance is a relevant Geophysics task, since in the early stages of seismic processing, we are required to identify and fix noisy sail lines. In this work, we introduce a real-world seismogram quality classification dataset based on 6,613 examples, manually labeled by human experts as good, bad or ugly, according to their noise intensity. This dataset is used to train a CNN classifier for seismic shot-gathers quality prediction. In our empirical evaluation, we observe an F1-score of 93.56% in the test set.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n A CNN-Based Tool to Index Emotion on Anime Character Stickers.\n \n \n \n\n\n \n Jesus, I.; Cardoso, J.; Busson, A.; Livio, Á.; Colcher, S.; and Milidiú, R.\n\n\n \n\n\n\n In pages 319–3193, December 2019. \n \n\n\n\n
\n\n\n\n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{jesus_cnn-based_2019,\n  title = {A {CNN}-{Based} {Tool} to {Index} {Emotion} on {Anime} {Character} {Stickers}},\n  doi = {10.1109/ISM46123.2019.00071},\n  author = {Jesus, Ivan and Cardoso, Jessica and Busson, Antonio and Livio, Álan and Colcher, Sérgio and Milidiú, Ruy},\n  month = dec,\n  year = {2019},\n  pages = {319--3193}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n PVBR-Recog: A YOLOv3-based Brazilian Automatic License Plate Recognition Tool.\n \n \n \n\n\n \n Ferreira Alves Pinto, P.; Busson, A.; Melo, J. P.; Colcher, S.; and Milidiú, R.\n\n\n \n\n\n\n In October 2019. \n \n\n\n\n
\n\n\n\n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ferreira_alves_pinto_pvbr-recog_2019,\n  title = {{PVBR}-{Recog}: {A} {YOLOv3}-based {Brazilian} {Automatic} {License} {Plate} {Recognition} {Tool}},\n  shorttitle = {{PVBR}-{Recog}},\n  doi = {10.5753/webmedia_estendido.2019.8149},\n  abstract = {Vehicle’s license plate detection and recognition is a task with several practical applications. It can be applied, for example, in the security segment, identifying stolen cars and controlling cars entry/exit in private areas. This work presents a Deep Learning based tool that uses the cascaded YOLOv3 to simultaneously detect and recognize vehicle plate. In experiments performed, our tool got a recall of 95\\% in plate detection and 96.2\\% accuracy in the recognition of the 7 characters of the license plate.},\n  author = {Ferreira Alves Pinto, Pedro and Busson, Antonio and Melo, João Paulo and Colcher, Sérgio and Milidiú, Ruy},\n  month = oct,\n  year = {2019}\n}\n\n
\n
\n\n\n
\n Vehicle’s license plate detection and recognition is a task with several practical applications. It can be applied, for example, in the security segment, identifying stolen cars and controlling cars entry/exit in private areas. This work presents a Deep Learning based tool that uses the cascaded YOLOv3 to simultaneously detect and recognize vehicle plate. In experiments performed, our tool got a recall of 95% in plate detection and 96.2% accuracy in the recognition of the 7 characters of the license plate.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Deep learning methods for video understanding.\n \n \n \n\n\n \n Noronha Pereira dos Santos, G.; Almeida de Freitas, P.; Busson, A.; Livio, Á.; Milidiú, R.; and Colcher, S.\n\n\n \n\n\n\n In pages 21–23, October 2019. \n \n\n\n\n
\n\n\n\n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{noronha_pereira_dos_santos_deep_2019,\n  title = {Deep learning methods for video understanding},\n  isbn = {978-1-4503-6763-9},\n  doi = {10.1145/3323503.3345029},\n  abstract = {Methods based on Deep Learning became state-of-the-art in several Multimedia challenges. However, there is a gap of professionals to perform Deep Learning in the industry. Therefore, this tutorial aims to present the grounds and ways to develop applications that uses methods based on Deep Learning for video analysis tasks. Likewise, this tutorial is an opportunity for students and information technology professionals to qualify themselves. The main focus of this short course is to present fundamentals and technologies to develop such methods of DL.},\n  author = {Noronha Pereira dos Santos, Gabriel and Almeida de Freitas, Pedro and Busson, Antonio and Livio, Álan and Milidiú, Ruy and Colcher, Sérgio},\n  month = oct,\n  year = {2019},\n  pages = {21--23}\n}\n\n
\n
\n\n\n
\n Methods based on Deep Learning became state-of-the-art in several Multimedia challenges. However, there is a gap of professionals to perform Deep Learning in the industry. Therefore, this tutorial aims to present the grounds and ways to develop applications that uses methods based on Deep Learning for video analysis tasks. Likewise, this tutorial is an opportunity for students and information technology professionals to qualify themselves. The main focus of this short course is to present fundamentals and technologies to develop such methods of DL.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Métodos baseados em Deep Learning para Análise de Vídeo.\n \n \n \n\n\n \n Noronha Pereira dos Santos, G.; Almeida de Freitas, P.; Busson, A.; Livio, Á.; Colcher, S.; and Milidiú, R.\n\n\n \n\n\n\n In pages 119. October 2019.\n \n\n\n\n
\n\n\n\n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@incollection{noronha_pereira_dos_santos_metodos_2019,\n  title = {Métodos baseados em {Deep} {Learning} para {Análise} de {Vídeo}},\n  isbn = {978-85-7669-481-6},\n  abstract = {Os métodos baseados no Deep Learning tornaram-se state-of-the-art em vários desafios de multimídia. No entanto, existe uma lacuna de profissionais para realizar o Deep Learning na indústria. Este capítulo tem como foco apresentar os fundamentos e tecnologias para desenvolver tais métodos de DL para analise de vídeo. Em especial, buscamos capacitar o leitor a: (1) entender os principais modelos baseados em DL, mais espe-cificamente Convolutional Neural Networks (CNN); (2) aplicar os modelos de DL para resolver tarefas de vídeo como: classificação de vídeo, classificação de multi-etiquetas de vídeo, detecção de objetos e estimação de pose. A linguagem de programação Python é apresentada em conjunto com a biblioteca TensorFlow para implementação dos modelos de DL.},\n  author = {Noronha Pereira dos Santos, Gabriel and Almeida de Freitas, Pedro and Busson, Antonio and Livio, Álan and Colcher, Sérgio and Milidiú, Ruy},\n  month = oct,\n  year = {2019},\n  doi = {10.5753/sbc.481.6.04},\n  pages = {119}\n}\n\n
\n
\n\n\n
\n Os métodos baseados no Deep Learning tornaram-se state-of-the-art em vários desafios de multimídia. No entanto, existe uma lacuna de profissionais para realizar o Deep Learning na indústria. Este capítulo tem como foco apresentar os fundamentos e tecnologias para desenvolver tais métodos de DL para analise de vídeo. Em especial, buscamos capacitar o leitor a: (1) entender os principais modelos baseados em DL, mais espe-cificamente Convolutional Neural Networks (CNN); (2) aplicar os modelos de DL para resolver tarefas de vídeo como: classificação de vídeo, classificação de multi-etiquetas de vídeo, detecção de objetos e estimação de pose. A linguagem de programação Python é apresentada em conjunto com a biblioteca TensorFlow para implementação dos modelos de DL.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Building a Massive Corpus for Named Entity Recognition using Free Open Data Sources Request PDF.\n \n \n \n \n\n\n \n \n\n\n \n\n\n\n 2019.\n \n\n\n\n
\n\n\n\n \n \n \"BuildingPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@misc{noauthor_building_nodate,\n  title = {Building a {Massive} {Corpus} for {Named} {Entity} {Recognition} using {Free} {Open} {Data} {Sources} {Request} {PDF}},\n  url = {https://www.researchgate.net/publication/335233220_Building_a_Massive_Corpus_for_Named_Entity_Recognition_using_Free_Open_Data_Sources},\n  abstract = {Request PDF Building a Massive Corpus for Named Entity Recognition using Free Open Data Sources With the recent progress in machine learning, boosted by techniques such as deep learning, many tasks can be successfully solved once a large... Find, read and cite all the research you need on ResearchGate},\n  language = {en},\n  year = {2019},\n  urldate = {2020-03-03},\n  journal = {ResearchGate}\n}\n\n
\n
\n\n\n
\n Request PDF Building a Massive Corpus for Named Entity Recognition using Free Open Data Sources With the recent progress in machine learning, boosted by techniques such as deep learning, many tasks can be successfully solved once a large... Find, read and cite all the research you need on ResearchGate\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n A Multimodal CNN-based Tool to Censure Inappropriate Video Scenes.\n \n \n \n\n\n \n Almeida de Freitas, P.; Mendes, P.; Noronha Pereira dos Santos, G.; Busson, A.; Livio, Á.; Colcher, S.; and Milidiú, R.\n\n\n \n\n\n\n November 2019.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@misc{almeida_de_freitas_multimodal_2019,\n  title = {A {Multimodal} {CNN}-based {Tool} to {Censure} {Inappropriate} {Video} {Scenes}},\n  abstract = {Due to the extensive use of video-sharing platforms and services for their storage, the amount of such media on the internet has become massive. This volume of data makes it difficult to control the kind of content that may be present in such video files. One of the main concerns regarding the video content is if it has an inappropriate subject matter, such as nudity, violence, or other potentially disturbing content. More than telling if a video is either appropriate or inappropriate, it is also important to identify which parts of it contain such content, for preserving parts that would be discarded in a simple broad analysis. In this work, we present a multimodal (using audio and image features) architecture based on Convolutional Neural Networks (CNNs) for detecting inappropriate scenes in video files. In the task of classifying video files, our model achieved 98.95\\% and 98.94\\% of F1-score for the appropriate and inappropriate classes, respectively. We also present a censoring tool that automatically censors inappropriate segments of a video file.},\n  author = {Almeida de Freitas, Pedro and Mendes, Paulo and Noronha Pereira dos Santos, Gabriel and Busson, Antonio and Livio, Álan and Colcher, Sérgio and Milidiú, Ruy},\n  month = nov,\n  year = {2019}\n}\n\n
\n
\n\n\n
\n Due to the extensive use of video-sharing platforms and services for their storage, the amount of such media on the internet has become massive. This volume of data makes it difficult to control the kind of content that may be present in such video files. One of the main concerns regarding the video content is if it has an inappropriate subject matter, such as nudity, violence, or other potentially disturbing content. More than telling if a video is either appropriate or inappropriate, it is also important to identify which parts of it contain such content, for preserving parts that would be discarded in a simple broad analysis. In this work, we present a multimodal (using audio and image features) architecture based on Convolutional Neural Networks (CNNs) for detecting inappropriate scenes in video files. In the task of classifying video files, our model achieved 98.95% and 98.94% of F1-score for the appropriate and inappropriate classes, respectively. We also present a censoring tool that automatically censors inappropriate segments of a video file.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Cumulative Sum Ranking.\n \n \n \n \n\n\n \n Milidiú, R. L.; and Rocha, R. H. S.\n\n\n \n\n\n\n November 2019.\n arXiv: 1911.11255\n\n\n\n
\n\n\n\n \n \n \"CumulativePaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@misc{milidiu_cumulative_2019,\n  title = {Cumulative {Sum} {Ranking}},\n  url = {http://arxiv.org/abs/1911.11255},\n  abstract = {The goal of Ordinal Regression is to find a rule that ranks items from a given set. Several learning algorithms to solve this prediction problem build an ensemble of binary classifiers. Ranking by Projecting uses interdependent binary perceptrons. These perceptrons share the same direction vector, but use different bias values. Similar approaches use independent direction vectors and biases. To combine the binary predictions, most of them adopt a simple counting heuristics. Here, we introduce a novel cumulative sum scoring function to combine the binary predictions. The proposed score value aggregates the strength of each one of the relevant binary classifications on how large is the item's rank. We show that our modeling casts ordinal regression as a Structured Perceptron problem. As a consequence, we simplify its formulation and description, which results in two simple online learning algorithms. The second algorithm is a Passive-Aggressive version of the first algorithm. We show that under some rank separability condition both algorithms converge. Furthermore, we provide mistake bounds for each one of the two online algorithms. For the Passive-Aggressive version, we assume the knowledge of a separation margin, what significantly improves the corresponding mistake bound. Additionally, we show that Ranking by Projecting is a special case of our prediction algorithm. From a neural network architecture point of view, our empirical findings suggest a layer of cusum units for ordinal regression, instead of the usual softmax layer of multiclass problems.},\n  urldate = {2020-03-03},\n  journal = {arXiv:1911.11255 [cs, stat]},\n  author = {Milidiú, Ruy Luiz and Rocha, Rafael Henrique Santos},\n  month = nov,\n  year = {2019},\n  note = {arXiv: 1911.11255}\n}
\n
\n\n\n
\n The goal of Ordinal Regression is to find a rule that ranks items from a given set. Several learning algorithms to solve this prediction problem build an ensemble of binary classifiers. Ranking by Projecting uses interdependent binary perceptrons. These perceptrons share the same direction vector, but use different bias values. Similar approaches use independent direction vectors and biases. To combine the binary predictions, most of them adopt a simple counting heuristics. Here, we introduce a novel cumulative sum scoring function to combine the binary predictions. The proposed score value aggregates the strength of each one of the relevant binary classifications on how large is the item's rank. We show that our modeling casts ordinal regression as a Structured Perceptron problem. As a consequence, we simplify its formulation and description, which results in two simple online learning algorithms. The second algorithm is a Passive-Aggressive version of the first algorithm. We show that under some rank separability condition both algorithms converge. Furthermore, we provide mistake bounds for each one of the two online algorithms. For the Passive-Aggressive version, we assume the knowledge of a separation margin, what significantly improves the corresponding mistake bound. Additionally, we show that Ranking by Projecting is a special case of our prediction algorithm. From a neural network architecture point of view, our empirical findings suggest a layer of cusum units for ordinal regression, instead of the usual softmax layer of multiclass problems.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2018\n \n \n (5)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Structured Prediction Networks through Latent Cost Learning.\n \n \n \n\n\n \n Milidiú, R.; and Rocha, R.\n\n\n \n\n\n\n In pages 645–649, November 2018. \n \n\n\n\n
\n\n\n\n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{milidiu_structured_2018,\n  title = {Structured {Prediction} {Networks} through {Latent} {Cost} {Learning}},\n  doi = {10.1109/SSCI.2018.8628625},\n  author = {Milidiú, Ruy and Rocha, Rafael},\n  month = nov,\n  year = {2018},\n  pages = {645--649}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Developing Deep Learning Models for Multimedia Applications in TensorFlow.\n \n \n \n\n\n \n Busson, A.; Figueiredo, L.; Noronha Pereira dos Santos, G.; Damasceno, A.; Colcher, S.; and Milidiú, R.\n\n\n \n\n\n\n In pages 7–9, October 2018. \n \n\n\n\n
\n\n\n\n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{busson_developing_2018,\n  title = {Developing {Deep} {Learning} {Models} for {Multimedia} {Applications} in {TensorFlow}},\n  doi = {10.1145/3243082.3264605},\n  abstract = {Methods based on Deep Learning became state-of-the-art in several Multimedia challenges. However, there is a gap of professionals to perform Deep Learning in the industry. Therefore, this short course aims to present the grounds and ways to develop multimedia applications using methods based on Deep Learning. Likewise, this short course is an opportunity for students and IT professionals can qualify yourselves.},\n  author = {Busson, Antonio and Figueiredo, Lucas and Noronha Pereira dos Santos, Gabriel and Damasceno, Andre and Colcher, Sérgio and Milidiú, Ruy},\n  month = oct,\n  year = {2018},\n  pages = {7--9}\n}\n\n
\n
\n\n\n
\n Methods based on Deep Learning became state-of-the-art in several Multimedia challenges. However, there is a gap of professionals to perform Deep Learning in the industry. Therefore, this short course aims to present the grounds and ways to develop multimedia applications using methods based on Deep Learning. Likewise, this short course is an opportunity for students and IT professionals can qualify yourselves.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n VideoRecognition - Uma Proposta de Serviço para Reconhecimento de Elementos de Vídeo em Larga Escala.\n \n \n \n\n\n \n Busson, A.; Livio, Á.; Noronha Pereira dos Santos, G.; Soares Neto, C.; Milidiú, R.; and Colcher, S.\n\n\n \n\n\n\n In October 2018. \n \n\n\n\n
\n\n\n\n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{busson_videorecognition_2018,\n  title = {{VideoRecognition} - {Uma} {Proposta} de {Serviço} para {Reconhecimento} de {Elementos} de {Vídeo} em {Larga} {Escala}},\n  abstract = {Deep Learning research has allowed significant advancement of various segments of multimedia, especially in tasks related to speech processing, hearing and computational vision. However, some video services are still focused only on the traditional use of media (capture, storage, transmission and presentation). In this paper, we discuss our ongoing research towards a DLaS, i.e. Deep Learning as a Service. This way, we present the state of art in video classification and recognition. Then we propose the VideoRecognition as DLaS to support the tasks such as: image classification and video scenes, object detection and facial recognition. We discuss the usage of the proposed service in the context of the video@RNP repository. Our main contributions consist on dicussussions over the state of art and it usage in nowdays multimedia services.},\n  author = {Busson, Antonio and Livio, Álan and Noronha Pereira dos Santos, Gabriel and Soares Neto, Carlos and Milidiú, Ruy and Colcher, Sérgio},\n  month = oct,\n  year = {2018},\n  doi = {10.13140/RG.2.2.32903.24481}\n}\n\n
\n
\n\n\n
\n Deep Learning research has allowed significant advancement of various segments of multimedia, especially in tasks related to speech processing, hearing and computational vision. However, some video services are still focused only on the traditional use of media (capture, storage, transmission and presentation). In this paper, we discuss our ongoing research towards a DLaS, i.e. Deep Learning as a Service. This way, we present the state of art in video classification and recognition. Then we propose the VideoRecognition as DLaS to support the tasks such as: image classification and video scenes, object detection and facial recognition. We discuss the usage of the proposed service in the context of the video@RNP repository. Our main contributions consist on dicussussions over the state of art and it usage in nowdays multimedia services.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Desenvolvendo Modelos de Deep Learning para Aplicações Multimídia no Tensorflow.\n \n \n \n\n\n \n Busson, A.; Figueiredo, L.; Noronha Pereira dos Santos, G.; Damasceno, A.; Colcher, S.; and Milidiú, R.\n\n\n \n\n\n\n In pages 67–116. October 2018.\n \n\n\n\n
\n\n\n\n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@incollection{busson_desenvolvendo_2018,\n  title = {Desenvolvendo {Modelos} de {Deep} {Learning} para {Aplicações} {Multimídia} no {Tensorflow}},\n  isbn = {978-85-7669-455-7},\n  abstract = {The availability of massive quantities of data, combined with increasing computational capabilities, makes it possible to develop more precise Machine Learning algorithms. These new tools provide advances in areas such as Natural Language Processing and Computer Vision, allowing efficient processing of images, text and audio. Now, cognitive functionalities, such as learning, recognition and detection, can be used in multimedia applications to create mechanisms beyond traditional capture, streaming and presenta- tion uses. Methods based on Deep Learning became state-of-the-art in several Multi- media challenges. This short course presents the grounds and ways to develop models using Deep Learning. It prepares the participant to: (1) understand and develop mod- els based on Deep Neural Networks, Convolutional Neural Networks (CNN), Recurrent Neural Networks, including LSTM and GRU; (2) apply the Deep Learning models to solve problems within the multimedia domain like Image Classification, Facial Recogni- tion, Object Detection, Video Scenes Classification. The Python programming language is shown alongside TensorFlow, a package for developing Deep Learning models.},\n  author = {Busson, Antonio and Figueiredo, Lucas and Noronha Pereira dos Santos, Gabriel and Damasceno, Andre and Colcher, Sérgio and Milidiú, Ruy},\n  month = oct,\n  year = {2018},\n  doi = {10.5753/sbc.455.7.03},\n  pages = {67--116}\n}\n\n
\n
\n\n\n
\n The availability of massive quantities of data, combined with increasing computational capabilities, makes it possible to develop more precise Machine Learning algorithms. These new tools provide advances in areas such as Natural Language Processing and Computer Vision, allowing efficient processing of images, text and audio. Now, cognitive functionalities, such as learning, recognition and detection, can be used in multimedia applications to create mechanisms beyond traditional capture, streaming and presenta- tion uses. Methods based on Deep Learning became state-of-the-art in several Multi- media challenges. This short course presents the grounds and ways to develop models using Deep Learning. It prepares the participant to: (1) understand and develop mod- els based on Deep Neural Networks, Convolutional Neural Networks (CNN), Recurrent Neural Networks, including LSTM and GRU; (2) apply the Deep Learning models to solve problems within the multimedia domain like Image Classification, Facial Recogni- tion, Object Detection, Video Scenes Classification. The Python programming language is shown alongside TensorFlow, a package for developing Deep Learning models.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Entropy-Guided Feature Generation for Large Margin Structured Learning.\n \n \n \n\n\n \n Fernandes, E.; and Milidiú, R.\n\n\n \n\n\n\n Monografias em Ciência da Computação. February 2018.\n \n\n\n\n
\n\n\n\n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{fernandes_entropy-guided_2018,\n  title = {Entropy-{Guided} {Feature} {Generation} for {Large} {Margin} {Structured} {Learning}},\n  doi = {10.17771/PUCRio.DImcc.24327},\n  journal = {Monografias em Ciência da Computação},\n  author = {Fernandes, Eraldo and Milidiú, Ruy},\n  month = feb,\n  year = {2018}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2014\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Latent Trees for Coreference Resolution.\n \n \n \n \n\n\n \n Fernandes, E. R.; dos Santos, C. N.; and Milidiú, R. L.\n\n\n \n\n\n\n Computational Linguistics, 40(4): 801–835. April 2014.\n \n\n\n\n
\n\n\n\n \n \n \"LatentPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{fernandes_latent_2014,\n  title = {Latent {Trees} for {Coreference} {Resolution}},\n  volume = {40},\n  issn = {0891-2017},\n  url = {https://www.mitpressjournals.org/doi/10.1162/COLI_a_00200},\n  doi = {10.1162/COLI_a_00200},\n  abstract = {We describe a structure learning system for unrestricted coreference resolution that explores two key modeling techniques: latent coreference trees and automatic entropy-guided feature induction. The latent tree modeling makes the learning problem computationally feasible because it incorporates a meaningful hidden structure. Additionally, using an automatic feature induction method, we can efficiently build enhanced nonlinear models using linear model learning algorithms. We present empirical results that highlight the contribution of each modeling technique used in the proposed system. Empirical evaluation is performed on the multilingual unrestricted coreference CoNLL-2012 Shared Task datasets, which comprise three languages: Arabic, Chinese and English. We apply the same system to all languages, except for minor adaptations to some language-dependent features such as nested mentions and specific static pronoun lists. A previous version of this system was submitted to the CoNLL-2012 Shared Task closed track, achieving an official score of 58.69, the best among the competitors. The unique enhancement added to the current system version is the inclusion of candidate arcs linking nested mentions for the Chinese language. By including such arcs, the score increases by almost 4.5 points for that language. The current system shows a score of 60.15, which corresponds to a 3.5\\% error reduction, and is the best performing system for each of the three languages.},\n  number = {4},\n  urldate = {2020-03-02},\n  journal = {Computational Linguistics},\n  author = {Fernandes, Eraldo Rezende and dos Santos, Cícero Nogueira and Milidiú, Ruy Luiz},\n  month = apr,\n  year = {2014},\n  pages = {801--835}\n}\n\n
\n
\n\n\n
\n We describe a structure learning system for unrestricted coreference resolution that explores two key modeling techniques: latent coreference trees and automatic entropy-guided feature induction. The latent tree modeling makes the learning problem computationally feasible because it incorporates a meaningful hidden structure. Additionally, using an automatic feature induction method, we can efficiently build enhanced nonlinear models using linear model learning algorithms. We present empirical results that highlight the contribution of each modeling technique used in the proposed system. Empirical evaluation is performed on the multilingual unrestricted coreference CoNLL-2012 Shared Task datasets, which comprise three languages: Arabic, Chinese and English. We apply the same system to all languages, except for minor adaptations to some language-dependent features such as nested mentions and specific static pronoun lists. A previous version of this system was submitted to the CoNLL-2012 Shared Task closed track, achieving an official score of 58.69, the best among the competitors. The unique enhancement added to the current system version is the inclusion of candidate arcs linking nested mentions for the Chinese language. By including such arcs, the score increases by almost 4.5 points for that language. The current system shows a score of 60.15, which corresponds to a 3.5% error reduction, and is the best performing system for each of the three languages.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2013\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Trading team composition for the intraday multistock market.\n \n \n \n \n\n\n \n Alvim, L. G. M.; and Milidiú, R. L.\n\n\n \n\n\n\n Decision Support Systems, 54(2): 838–845. January 2013.\n \n\n\n\n
\n\n\n\n \n \n \"TradingPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{alvim_trading_2013,\n  title = {Trading team composition for the intraday multistock market},\n  volume = {54},\n  issn = {0167-9236},\n  url = {http://www.sciencedirect.com/science/article/pii/S0167923612002448},\n  doi = {10.1016/j.dss.2012.09.009},\n  abstract = {Automated traders operate market shares without human intervention. We propose a Trading Team based on atomic traders with opportunity detectors and simple effectors. The detectors signalize trading opportunities. For each trading signal, the effectors follow deterministic rules on when and what to trade in the market. The detectors are based on Partial Least Squares. We perform some trading experiments with twelve BM\\&FBovespa stocks. The empirical findings indicate that the proposed trading strategy reaches a 77.26\\% annualized profit, outperforming by 380.07\\% the chosen baseline strategy with a 16.07\\% profit. We also investigate Multistock Resolution Strategy (MSR) performance subject to brokerage commissions and income tax. Whenever the initial investment is at least US\\$ 50,000, the MSR strategy provides a profit of at least 38.63\\%.},\n  language = {en},\n  number = {2},\n  urldate = {2020-03-02},\n  journal = {Decision Support Systems},\n  author = {Alvim, Leandro G. M. and Milidiú, Ruy L.},\n  month = jan,\n  year = {2013},\n  pages = {838--845}\n}\n\n
\n
\n\n\n
\n Automated traders operate market shares without human intervention. We propose a Trading Team based on atomic traders with opportunity detectors and simple effectors. The detectors signalize trading opportunities. For each trading signal, the effectors follow deterministic rules on when and what to trade in the market. The detectors are based on Partial Least Squares. We perform some trading experiments with twelve BM&FBovespa stocks. The empirical findings indicate that the proposed trading strategy reaches a 77.26% annualized profit, outperforming by 380.07% the chosen baseline strategy with a 16.07% profit. We also investigate Multistock Resolution Strategy (MSR) performance subject to brokerage commissions and income tax. Whenever the initial investment is at least US$ 50,000, the MSR strategy provides a profit of at least 38.63%.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2012\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Double dip map-reduce for processing cross validation jobs.\n \n \n \n \n\n\n \n Moret, D.; Breitman, K.; Amorim, E.; Talavera, J.; Milidiu, R.; and Viterbo, J.\n\n\n \n\n\n\n In Proceedings of the 27th Annual ACM Symposium on Applied Computing, of SAC '12, pages 473–477, Trento, Italy, March 2012. Association for Computing Machinery\n \n\n\n\n
\n\n\n\n \n \n \"DoublePaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{moret_double_2012,\n  address = {Trento, Italy},\n  series = {{SAC} '12},\n  title = {Double dip map-reduce for processing cross validation jobs},\n  isbn = {978-1-4503-0857-1},\n  url = {https://doi.org/10.1145/2245276.2245367},\n  doi = {10.1145/2245276.2245367},\n  abstract = {Cross validation is fundamental to machine learning as it provides a reliable way in which to evaluate algorithms and the overall quality of the corpora in use. In typical cross validation, the corpus is initially divided into learning and training segments, then crossed-over in successive rounds, so that each data segment is validated against the remaining ones. This process is prohibitively time and effort consuming, and often brushed off for computationally cheaper ones, such as heuristics. In this paper we introduce a cloud-based architecture for running cross validation jobs. Our solution makes heavy use of computational resources in the cloud by proposing a strategy in which there are two distinct, subsequent, map-reduce cycles: the first to perform the algorithmic target computation, and the second to provide cross validation data to retrofit the machine learning process. We demonstrate the feasibility of the proposed approach, with the implementation of a web segmentation algorithm.},\n  urldate = {2020-03-02},\n  booktitle = {Proceedings of the 27th {Annual} {ACM} {Symposium} on {Applied} {Computing}},\n  publisher = {Association for Computing Machinery},\n  author = {Moret, Danilo and Breitman, Karin and Amorim, Evelin and Talavera, Jose and Milidiu, Ruy and Viterbo, Jose},\n  month = mar,\n  year = {2012},\n  pages = {473--477}\n}\n\n
\n
\n\n\n
\n Cross validation is fundamental to machine learning as it provides a reliable way in which to evaluate algorithms and the overall quality of the corpora in use. In typical cross validation, the corpus is initially divided into learning and training segments, then crossed-over in successive rounds, so that each data segment is validated against the remaining ones. This process is prohibitively time and effort consuming, and often brushed off for computationally cheaper ones, such as heuristics. In this paper we introduce a cloud-based architecture for running cross validation jobs. Our solution makes heavy use of computational resources in the cloud by proposing a strategy in which there are two distinct, subsequent, map-reduce cycles: the first to perform the algorithmic target computation, and the second to provide cross validation data to retrofit the machine learning process. We demonstrate the feasibility of the proposed approach, with the implementation of a web segmentation algorithm.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2011\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Quotation extraction for portuguese.\n \n \n \n\n\n \n Fernandes, W. P. D.; Motta, E.; and Milidiú, R. L.\n\n\n \n\n\n\n In Proceedings of the 8th Brazilian Symposium in Information and Human Language Technology, 2011. \n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{fernandes_quotation_2011,\n  title = {Quotation extraction for portuguese},\n  booktitle = {Proceedings of the 8th {Brazilian} {Symposium} in {Information} and {Human} {Language} {Technology}},\n  author = {Fernandes, William Paulo Ducca and Motta, Eduardo and Milidiú, Ruy Luiz},\n  year = {2011}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2010\n \n \n (3)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n RelHunter : a machine learning method for relation extraction from text.\n \n \n \n \n\n\n \n Fernandes, E. R.; Milidiú, R. L.; and Rentería, R. P.\n\n\n \n\n\n\n Journal of the Brazilian Computer Society, 16(3): 191–199. September 2010.\n \n\n\n\n
\n\n\n\n \n \n \"RelHunterPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{fernandes_relhunter_2010,\n  title = {{RelHunter} : a machine learning method for relation extraction from text},\n  volume = {16},\n  copyright = {2010 The Brazilian Computer Society},\n  issn = {1678-4804},\n  shorttitle = {{RelHunter}},\n  url = {https://journal-bcs.springeropen.com/articles/10.1007/s13173-010-0018-y},\n  doi = {10.1007/s13173-010-0018-y},\n  abstract = {We propose RelHunter, a machine learning-based method for the extraction of structured information from text. RelHunter’s key idea is to model the target structures as a relation over entities. Hence, the modeling effort is reduced to the identification of entities and the generation of a candidate relation, which are simpler problems than the original one. RelHunter fits a very broad spectrum of complex computational linguistic problems. We apply it to five tasks: phrase chunking, clause identification, hedge detection, quotation extraction, and dependency parsing. We compare RelHunter to token classification approaches through several computational experiments on seven multilingual corpora. RelHunter outperforms the token classification approaches by 2.14\\% on average. Moreover, we compare the derived systems against state-of-the-art systems for each corpus. Our systems achieve state-of-the-art performances for three corpora: Portuguese phrase chunking, Portuguese clause identification, and English quotation extraction. Additionally, the derived systems show good quality performance for the other four corpora.},\n  language = {en},\n  number = {3},\n  urldate = {2020-03-02},\n  journal = {Journal of the Brazilian Computer Society},\n  author = {Fernandes, Eraldo R. and Milidiú, Ruy L. and Rentería, Raúl P.},\n  month = sep,\n  year = {2010},\n  pages = {191--199}\n}\n\n
\n
\n\n\n
\n We propose RelHunter, a machine learning-based method for the extraction of structured information from text. RelHunter’s key idea is to model the target structures as a relation over entities. Hence, the modeling effort is reduced to the identification of entities and the generation of a candidate relation, which are simpler problems than the original one. RelHunter fits a very broad spectrum of complex computational linguistic problems. We apply it to five tasks: phrase chunking, clause identification, hedge detection, quotation extraction, and dependency parsing. We compare RelHunter to token classification approaches through several computational experiments on seven multilingual corpora. RelHunter outperforms the token classification approaches by 2.14% on average. Moreover, we compare the derived systems against state-of-the-art systems for each corpus. Our systems achieve state-of-the-art performances for three corpora: Portuguese phrase chunking, Portuguese clause identification, and English quotation extraction. Additionally, the derived systems show good quality performance for the other four corpora.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Data stream anomaly detection through principal subspace tracking.\n \n \n \n\n\n \n dos Santos Teixeira, P. H.; and Milidiú, R. L.\n\n\n \n\n\n\n In Proceedings of the 2010 ACM Symposium on Applied Computing, pages 1609–1616, 2010. \n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{dos_santos_teixeira_data_2010,\n  title = {Data stream anomaly detection through principal subspace tracking},\n  booktitle = {Proceedings of the 2010 {ACM} {Symposium} on {Applied} {Computing}},\n  author = {dos Santos Teixeira, Pedro Henriques and Milidiú, Ruy Luiz},\n  year = {2010},\n  pages = {1609--1616}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Daily volume forecasting using high frequency predictors.\n \n \n \n\n\n \n Alvim, L. G.; dos Santos, C. N.; and Milidiu, R. L.\n\n\n \n\n\n\n In Proceedings of the 10th IASTED International Conference, volume 674, pages 248, 2010. \n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{alvim_daily_2010,\n  title = {Daily volume forecasting using high frequency predictors},\n  volume = {674},\n  booktitle = {Proceedings of the 10th {IASTED} {International} {Conference}},\n  author = {Alvim, Leandro GM and dos Santos, Cicero N. and Milidiu, Ruy L.},\n  year = {2010},\n  pages = {248}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2009\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Entropy guided transformation learning.\n \n \n \n\n\n \n dos Santos, C. N.; and Milidiú, R. L.\n\n\n \n\n\n\n In Foundations of Computational, Intelligence Volume 1, pages 159–184. Springer, 2009.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@incollection{dos_santos_entropy_2009,\n  title = {Entropy guided transformation learning},\n  booktitle = {Foundations of {Computational}, {Intelligence} {Volume} 1},\n  publisher = {Springer},\n  author = {dos Santos, Cícero Nogueira and Milidiú, Ruy Luiz},\n  year = {2009},\n  pages = {159--184}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2008\n \n \n (3)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Portuguese corpus-based learning using ETL.\n \n \n \n \n\n\n \n Milidiú, R. L.; Santos, C. N. d.; and Duarte, J. C.\n\n\n \n\n\n\n Journal of the Brazilian Computer Society, 14(4): 17–27. December 2008.\n \n\n\n\n
\n\n\n\n \n \n \"PortuguesePaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{milidiu_portuguese_2008,\n  title = {Portuguese corpus-based learning using {ETL}},\n  volume = {14},\n  issn = {0104-6500},\n  url = {http://www.scielo.br/scielo.php?script=sci_abstract&pid=S0104-65002008000400003&lng=en&nrm=iso&tlng=en},\n  doi = {10.1007/BF03192569},\n  language = {en},\n  number = {4},\n  urldate = {2020-03-02},\n  journal = {Journal of the Brazilian Computer Society},\n  author = {Milidiú, Ruy Luiz and Santos, Cícero Nogueira dos and Duarte, Julio Cesar},\n  month = dec,\n  year = {2008},\n  pages = {17--27}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Instance-Based Ontology Mapping.\n \n \n \n\n\n \n Breitman, K. K.; Brauner, D.; Casanova, M. A.; Milidiú, R.; Gazola, A.; and Perazolo, M.\n\n\n \n\n\n\n In Fifth IEEE Workshop on Engineering of Autonomic and Autonomous Systems (ease 2008), pages 67–74, March 2008. \n ISSN: 2168-1872\n\n\n\n
\n\n\n\n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{breitman_instance-based_2008,\n  title = {Instance-{Based} {Ontology} {Mapping}},\n  doi = {10.1109/EASe.2008.18},\n  abstract = {We have explored the use of formal ontology and alignment techniques as a general approach to mediate and reconcile different representations. This approach proved to be very effective when fueled by ontologies rich in detail, (properties, restrictions, attributes and axioms that hold among classes), but performs poorly when available representations are incomplete or lacking, which is often the case in real life settings. In this paper we propose an enhancement to the original approach, based on a, instance-based matching technique.},\n  booktitle = {Fifth {IEEE} {Workshop} on {Engineering} of {Autonomic} and {Autonomous} {Systems} (ease 2008)},\n  author = {Breitman, Karin K. and Brauner, Daniela and Casanova, Marco Antonio and Milidiú, Ruy and Gazola, Alexandre and Perazolo, Marcelo},\n  month = mar,\n  year = {2008},\n  note = {ISSN: 2168-1872},\n  pages = {67--74}\n}\n\n
\n
\n\n\n
\n We have explored the use of formal ontology and alignment techniques as a general approach to mediate and reconcile different representations. This approach proved to be very effective when fueled by ontologies rich in detail, (properties, restrictions, attributes and axioms that hold among classes), but performs poorly when available representations are incomplete or lacking, which is often the case in real life settings. In this paper we propose an enhancement to the original approach, based on a, instance-based matching technique.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Phrase chunking using entropy guided transformation learning.\n \n \n \n\n\n \n Milidiú, R. L.; dos Santos, C.; and Duarte, J. C.\n\n\n \n\n\n\n In Proceedings of ACL-08: HLT, pages 647–655, 2008. \n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{milidiu_phrase_2008,\n  title = {Phrase chunking using entropy guided transformation learning},\n  booktitle = {Proceedings of {ACL}-08: {HLT}},\n  author = {Milidiú, Ruy Luiz and dos Santos, Cicero and Duarte, Julio C.},\n  year = {2008},\n  pages = {647--655}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2007\n \n \n (2)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n MKPLS approach: switching strategies for the non-linear multi-kernel PLSR.\n \n \n \n \n\n\n \n Rentería, R.; Milidiú, R.; and Souza, R.\n\n\n \n\n\n\n Computational Statistics, 22(2): 323–330. July 2007.\n \n\n\n\n
\n\n\n\n \n \n \"MKPLSPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{renteria_mkpls_2007,\n  title = {{MKPLS} approach: switching strategies for the non-linear multi-kernel {PLSR}},\n  volume = {22},\n  issn = {1613-9658},\n  shorttitle = {{MKPLS} approach},\n  url = {https://doi.org/10.1007/s00180-007-0040-5},\n  doi = {10.1007/s00180-007-0040-5},\n  abstract = {We present two strategies to determine the kernel switching order for the non-linear multi-kernel PLSR algorithm. The multi-kernel PLS (MKPLS) algorithm builds upon a one kernel PLSR which uses a kernel matrix to hold the inner products of the projection of the independent data set onto a feature space. After building a PLSR model, MKPLS deflates the kernel matrix so that only that part which cannot be predicted by the model remains. This remainder is projected onto a different feature space and a new PLSR model is built. The switching algorithms presented for this approach address two questions: which kernel should be used at each iteration and; how many factors should be extracted before switching to another kernel.},\n  language = {en},\n  number = {2},\n  urldate = {2020-03-02},\n  journal = {Computational Statistics},\n  author = {Rentería, Raúl and Milidiú, Ruy and Souza, Rafael},\n  month = jul,\n  year = {2007},\n  pages = {323--330}\n}\n\n
\n
\n\n\n
\n We present two strategies to determine the kernel switching order for the non-linear multi-kernel PLSR algorithm. The multi-kernel PLS (MKPLS) algorithm builds upon a one kernel PLSR which uses a kernel matrix to hold the inner products of the projection of the independent data set onto a feature space. After building a PLSR model, MKPLS deflates the kernel matrix so that only that part which cannot be predicted by the model remains. This remainder is projected onto a different feature space and a new PLSR model is built. The switching algorithms presented for this approach address two questions: which kernel should be used at each iteration and; how many factors should be extracted before switching to another kernel.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n TBL Template Selection: An Evolutionary Approach.\n \n \n \n\n\n \n Milidiú, R. L.; Duarte, J. C.; and Nogueira dos Santos, C.\n\n\n \n\n\n\n In Borrajo, D.; Castillo, L.; and Corchado, J. M., editor(s), Current Topics in Artificial Intelligence, of Lecture Notes in Computer Science, pages 180–189, Berlin, Heidelberg, 2007. Springer\n \n\n\n\n
\n\n\n\n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{milidiu_tbl_2007,\n  address = {Berlin, Heidelberg},\n  series = {Lecture {Notes} in {Computer} {Science}},\n  title = {{TBL} {Template} {Selection}: {An} {Evolutionary} {Approach}},\n  isbn = {978-3-540-75271-4},\n  shorttitle = {{TBL} {Template} {Selection}},\n  doi = {10.1007/978-3-540-75271-4_19},\n  abstract = {Transformation Based Learning (TBL) is an intensively Machine Learning algorithm frequently used in Natural Language Processing. TBL uses rule templates to identify error-correcting patterns. A critical requirement in TBL is the availability of a problem domain expert to build these rule templates. In this work, we propose an evolutionary approach based on Genetic Algorithms to automatically implement the template selection process. We show some empirical evidence that our approach provides template sets with almost the same quality as human built templates.},\n  language = {en},\n  booktitle = {Current {Topics} in {Artificial} {Intelligence}},\n  publisher = {Springer},\n  author = {Milidiú, Ruy Luiz and Duarte, Julio Cesar and Nogueira dos Santos, Cícero},\n  editor = {Borrajo, Daniel and Castillo, Luis and Corchado, Juan Manuel},\n  year = {2007},\n  pages = {180--189}\n}\n\n
\n
\n\n\n
\n Transformation Based Learning (TBL) is an intensively Machine Learning algorithm frequently used in Natural Language Processing. TBL uses rule templates to identify error-correcting patterns. A critical requirement in TBL is the availability of a problem domain expert to build these rule templates. In this work, we propose an evolutionary approach based on Genetic Algorithms to automatically implement the template selection process. We show some empirical evidence that our approach provides template sets with almost the same quality as human built templates.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2005\n \n \n (2)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n An Agent Based Architecture for Highly Competitive Electronic Markets.\n \n \n \n\n\n \n Sardinha, J. A. R.; Milidiú, R. L.; Paranhos, P. M.; Cunha, P. M.; and de Lucena, C. J. P.\n\n\n \n\n\n\n In FLAIRS Conference, pages 326–332, 2005. \n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{sardinha_agent_2005,\n  title = {An {Agent} {Based} {Architecture} for {Highly} {Competitive} {Electronic} {Markets}.},\n  booktitle = {{FLAIRS} {Conference}},\n  author = {Sardinha, José Alberto RP and Milidiú, Ruy Luiz and Paranhos, Patrick M. and Cunha, Pedro M. and de Lucena, Carlos José Pereira},\n  year = {2005},\n  pages = {326--332}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n DPLS and PPLS: two PLS algorithms for large data sets.\n \n \n \n\n\n \n Milidiu, R. L.; and Renterı́a, R. P.\n\n\n \n\n\n\n Computational statistics & data analysis, 48(1): 125–138. 2005.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{milidiu_dpls_2005,\n  title = {{DPLS} and {PPLS}: two {PLS} algorithms for large data sets},\n  volume = {48},\n  shorttitle = {{DPLS} and {PPLS}},\n  number = {1},\n  journal = {Computational statistics \\& data analysis},\n  author = {Milidiu, Ruy L. and Renterı́a, Raúl P.},\n  year = {2005},\n  pages = {125--138}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2003\n \n \n (2)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n An Object-Oriented Framework for Building Software Agents.\n \n \n \n\n\n \n Sardinha, J. A. R. P.; Ribeiro, P. C.; Milidiú, R. L.; and de Lucena, C. J. P.\n\n\n \n\n\n\n Journal of Object Technology, 2(1): 85–97. 2003.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{sardinha_object-oriented_2003,\n  title = {An {Object}-{Oriented} {Framework} for {Building} {Software} {Agents}.},\n  volume = {2},\n  number = {1},\n  journal = {Journal of Object Technology},\n  author = {Sardinha, José Alberto Rodrigues Pereira and Ribeiro, Paula Clark and Milidiú, Ruy Luiz and de Lucena, Carlos José Pereira},\n  year = {2003},\n  pages = {85--97}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n The complexity of makespan minimization for pipeline transportation.\n \n \n \n\n\n \n Milidiú, R. L.; Pessoa, A. A.; and Laber, E. S.\n\n\n \n\n\n\n Theoretical Computer Science, 306(1-3): 339–351. 2003.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{milidiu_complexity_2003,\n  title = {The complexity of makespan minimization for pipeline transportation},\n  volume = {306},\n  number = {1-3},\n  journal = {Theoretical Computer Science},\n  author = {Milidiú, Ruy Luiz and Pessoa, Artur Alves and Laber, Eduardo Sany},\n  year = {2003},\n  pages = {339--351}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2001\n \n \n (2)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Bounding the inefficiency of length-restricted prefix codes.\n \n \n \n\n\n \n Milidiú, R. L.; and Laber, E. S.\n\n\n \n\n\n\n Algorithmica, 31(4): 513–529. 2001.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{milidiu_bounding_2001,\n  title = {Bounding the inefficiency of length-restricted prefix codes},\n  volume = {31},\n  number = {4},\n  journal = {Algorithmica},\n  author = {Milidiú, Ruy Luiz and Laber, Eduardo Sany},\n  year = {2001},\n  pages = {513--529}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Three space-economical algorithms for calculating minimum-redundancy prefix codes.\n \n \n \n\n\n \n Milidiú, R. L.; Pessoa, A. A.; and Laber, E. S.\n\n\n \n\n\n\n IEEE Transactions on Information Theory, 47(6): 2185–2198. 2001.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{milidiu_three_2001,\n  title = {Three space-economical algorithms for calculating minimum-redundancy prefix codes},\n  volume = {47},\n  number = {6},\n  journal = {IEEE Transactions on Information Theory},\n  author = {Milidiú, Ruy Luiz and Pessoa, Artur Alves and Laber, Eduardo Sany},\n  year = {2001},\n  pages = {2185--2198}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2000\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n The warm-up algorithm: a Lagrangian construction of length restricted Huffman codes.\n \n \n \n\n\n \n Milidiú, R. L.; and Laber, E. S.\n\n\n \n\n\n\n SIAM Journal on Computing, 30(5): 1405–1426. 2000.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{milidiu_warm-up_2000,\n  title = {The warm-up algorithm: a {Lagrangian} construction of length restricted {Huffman} codes},\n  volume = {30},\n  shorttitle = {The warm-up algorithm},\n  number = {5},\n  journal = {SIAM Journal on Computing},\n  author = {Milidiú, Ruy Luiz and Laber, Eduardo Sany},\n  year = {2000},\n  pages = {1405--1426}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 1999\n \n \n (3)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Time-series forecasting through wavelets transformation and a mixture of expert models.\n \n \n \n\n\n \n Milidiú, R. L.; Machado, R. J.; and Renterı́a, R. P.\n\n\n \n\n\n\n Neurocomputing, 28(1-3): 145–156. 1999.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{milidiu_time-series_1999,\n  title = {Time-series forecasting through wavelets transformation and a mixture of expert models},\n  volume = {28},\n  number = {1-3},\n  journal = {Neurocomputing},\n  author = {Milidiú, Ruy L. and Machado, Ricardo J. and Renterı́a, Raúl P.},\n  year = {1999},\n  pages = {145--156}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Efficient implementation of the WARM-UP algorithm for the construction of length-restricted prefix codes.\n \n \n \n\n\n \n Milidiú, R. L.; Pessoa, A. A.; and Laber, E. S.\n\n\n \n\n\n\n In Workshop on Algorithm Engineering and Experimentation, pages 1–17, 1999. Springer\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{milidiu_efficient_1999,\n  title = {Efficient implementation of the {WARM}-{UP} algorithm for the construction of length-restricted prefix codes},\n  booktitle = {Workshop on {Algorithm} {Engineering} and {Experimentation}},\n  publisher = {Springer},\n  author = {Milidiú, Ruy Luiz and Pessoa, Artur Alves and Laber, Eduardo Sany},\n  year = {1999},\n  pages = {1--17}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Bounding the compression loss of the FGK algorithm.\n \n \n \n\n\n \n Milidiú, R. L.; Laber, E. S.; and Pessoa, A. A.\n\n\n \n\n\n\n Journal of Algorithms, 32(2): 195–211. 1999.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{milidiu_bounding_1999,\n  title = {Bounding the compression loss of the {FGK} algorithm},\n  volume = {32},\n  number = {2},\n  journal = {Journal of Algorithms},\n  author = {Milidiú, Ruy Luiz and Laber, Eduardo Sany and Pessoa, Artur Alves},\n  year = {1999},\n  pages = {195--211}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 1993\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Belief function model for information retrieval.\n \n \n \n\n\n \n da Silva, W. T.; and Milidiú, R. L.\n\n\n \n\n\n\n Journal of the American Society for Information Science, 44(1): 10–18. 1993.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{da_silva_belief_1993,\n  title = {Belief function model for information retrieval},\n  volume = {44},\n  number = {1},\n  journal = {Journal of the American Society for Information Science},\n  author = {da Silva, Wagner Teixeira and Milidiú, Ruy Luiz},\n  year = {1993},\n  pages = {10--18}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 1992\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Algorithms for combining belief functions.\n \n \n \n\n\n \n da Silva, W. T.; and Milidiu, R. L.\n\n\n \n\n\n\n International Journal of Approximate Reasoning, 7(1-2): 73–94. 1992.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{da_silva_algorithms_1992,\n  title = {Algorithms for combining belief functions},\n  volume = {7},\n  number = {1-2},\n  journal = {International Journal of Approximate Reasoning},\n  author = {da Silva, Wagner Teixeira and Milidiu, Ruy Luiz},\n  year = {1992},\n  pages = {73--94}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 1987\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n The Computation Of Compound Distributions With Two-Sided Severities (Negative-Binomial Iterative, Recursive, Asymptotic, Sum).\n \n \n \n\n\n \n Milidiu, R. L.\n\n\n \n\n\n\n . 1987.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{milidiu_computation_1987,\n  title = {The {Computation} {Of} {Compound} {Distributions} {With} {Two}-{Sided} {Severities} ({Negative}-{Binomial} {Iterative}, {Recursive}, {Asymptotic}, {Sum})},\n  author = {Milidiu, Ruy Luiz},\n  year = {1987}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n\n\n\n
\n\n\n \n\n \n \n \n \n\n
\n"}; document.write(bibbase_data.data);