Automatically estimating emotion in music with deep long-short term memory recurrent neural networks. Coutinho, E., Trigeorgis, G., Zafeiriou, S., & Schuller, B. In Larson, M., Ionescu, B., Sjöberg, M., Anguera, X., Poignant, J., Riegler, M., Eskevich, M., Hauff, C., Sutcliffe, R., Jones, G., J., Yang, Y., Soleymani, M., & Papadopoulos, S., editors, CEUR Workshop Proceedings, volume 1436, pages 1-3, 9, 2015. CEUR. Paper Website abstract bibtex 1 download In this paper we describe our approach for the MediaEval's "Emotion in Music" task. Our method consists of deep Long-Short Term Memory Recurrent Neural Networks (LSTM-RNN) for dynamic Arousal and Valence regression, using acoustic and psychoacoustic features extracted from the songs that have been previously proven as effective for emotion prediction in music. Results on the challenge test demonstrate an excellent performance for Arousal estimation (r = 0.613 ± 0.278), but not for Valence (r = 0.026 ± 0.500). Issues regarding the quality of the test set annotations' reliability and distributions are indicated as plausible justifications for these results. By using a subset of the development set that was left out for performance estimation, we could determine that the performance of our approach may be underestimated for Valence (Arousal: r = 0.596 ± 0.386; Valence: r = 0.458 ± 0.551).
@inproceedings{
title = {Automatically estimating emotion in music with deep long-short term memory recurrent neural networks},
type = {inproceedings},
year = {2015},
pages = {1-3},
volume = {1436},
websites = {https://www.scopus.com/inward/record.uri?eid=2-s2.0-84989923198&partnerID=40&md5=f0d407c8ffb96c19f299c78435558371,http://ceur-ws.org/Vol-1436/Paper64.pdf},
month = {9},
publisher = {CEUR},
city = {Wurzen, Germany},
id = {d43d00f9-3521-3815-bd81-243f874a33a5},
created = {2024-08-09T12:19:57.621Z},
file_attached = {true},
profile_id = {ffa9027c-806a-3827-93a1-02c42eb146a1},
group_id = {da2a8249-fdf4-3036-ba56-7358198a1600},
last_modified = {2024-08-09T12:20:58.011Z},
read = {false},
starred = {false},
authored = {false},
confirmed = {true},
hidden = {false},
citation_key = {coutinho2015automaticallynetworks},
source_type = {inproceedings},
notes = {<b>From Duplicate 1 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, Eduardo; Trigeorgis, George; Zafeiriou, Stefanos; Schuller, Björn)<br/></b><br/><b>From Duplicate 1 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, Eduardo; Trigeorgis, George; Zafeiriou, Stefanos; Schuller, Björn)<br/></b><br/>cited By 7<br/><br/><b>From Duplicate 2 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, Eduardo; Trigeorgis, George; Zafeiriou, Stefanos; Schuller, Björn)<br/></b><br/><b>From Duplicate 1 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, E; Trigeorgis, G; Zafeiriou, S; Schuller, B)<br/></b><br/>cited By 7<br/><br/><b>From Duplicate 2 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, Eduardo; Trigeorgis, George; Zafeiriou, Stefanos; Schuller, Björn)<br/></b><br/><b>From Duplicate 1 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, E; Trigeorgis, G; Zafeiriou, S; Schuller, B)<br/></b><br/>cited By 7},
private_publication = {false},
abstract = {In this paper we describe our approach for the MediaEval's "Emotion in Music" task. Our method consists of deep Long-Short Term Memory Recurrent Neural Networks (LSTM-RNN) for dynamic Arousal and Valence regression, using acoustic and psychoacoustic features extracted from the songs that have been previously proven as effective for emotion prediction in music. Results on the challenge test demonstrate an excellent performance for Arousal estimation (r = 0.613 ± 0.278), but not for Valence (r = 0.026 ± 0.500). Issues regarding the quality of the test set annotations' reliability and distributions are indicated as plausible justifications for these results. By using a subset of the development set that was left out for performance estimation, we could determine that the performance of our approach may be underestimated for Valence (Arousal: r = 0.596 ± 0.386; Valence: r = 0.458 ± 0.551).},
bibtype = {inproceedings},
author = {Coutinho, Eduardo and Trigeorgis, George and Zafeiriou, Stefanos and Schuller, Björn},
editor = {Larson, Martha and Ionescu, Bogdan and Sjöberg, Mats and Anguera, Xavier and Poignant, Johann and Riegler, Michael and Eskevich, Maria and Hauff, Claudia and Sutcliffe, Richard and Jones, Gareth J.F. and Yang, Yi-Hsuan and Soleymani, Mohammad and Papadopoulos, Symeon},
booktitle = {CEUR Workshop Proceedings},
keywords = {article,conference}
}
Downloads: 1
{"_id":"oLv7yej58ur7pcBYb","bibbaseid":"coutinho-trigeorgis-zafeiriou-schuller-automaticallyestimatingemotioninmusicwithdeeplongshorttermmemoryrecurrentneuralnetworks-2015","downloads":1,"creationDate":"2015-10-23T23:02:01.594Z","title":"Automatically estimating emotion in music with deep long-short term memory recurrent neural networks","author_short":["Coutinho, E.","Trigeorgis, G.","Zafeiriou, S.","Schuller, B."],"year":2015,"bibtype":"inproceedings","biburl":"https://bibbase.org/service/mendeley/ffa9027c-806a-3827-93a1-02c42eb146a1","bibdata":{"title":"Automatically estimating emotion in music with deep long-short term memory recurrent neural networks","type":"inproceedings","year":"2015","pages":"1-3","volume":"1436","websites":"https://www.scopus.com/inward/record.uri?eid=2-s2.0-84989923198&partnerID=40&md5=f0d407c8ffb96c19f299c78435558371,http://ceur-ws.org/Vol-1436/Paper64.pdf","month":"9","publisher":"CEUR","city":"Wurzen, Germany","id":"d43d00f9-3521-3815-bd81-243f874a33a5","created":"2024-08-09T12:19:57.621Z","file_attached":"true","profile_id":"ffa9027c-806a-3827-93a1-02c42eb146a1","group_id":"da2a8249-fdf4-3036-ba56-7358198a1600","last_modified":"2024-08-09T12:20:58.011Z","read":false,"starred":false,"authored":false,"confirmed":"true","hidden":false,"citation_key":"coutinho2015automaticallynetworks","source_type":"inproceedings","notes":"<b>From Duplicate 1 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, Eduardo; Trigeorgis, George; Zafeiriou, Stefanos; Schuller, Björn)<br/></b><br/><b>From Duplicate 1 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, Eduardo; Trigeorgis, George; Zafeiriou, Stefanos; Schuller, Björn)<br/></b><br/>cited By 7<br/><br/><b>From Duplicate 2 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, Eduardo; Trigeorgis, George; Zafeiriou, Stefanos; Schuller, Björn)<br/></b><br/><b>From Duplicate 1 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, E; Trigeorgis, G; Zafeiriou, S; Schuller, B)<br/></b><br/>cited By 7<br/><br/><b>From Duplicate 2 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, Eduardo; Trigeorgis, George; Zafeiriou, Stefanos; Schuller, Björn)<br/></b><br/><b>From Duplicate 1 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, E; Trigeorgis, G; Zafeiriou, S; Schuller, B)<br/></b><br/>cited By 7","private_publication":false,"abstract":"In this paper we describe our approach for the MediaEval's \"Emotion in Music\" task. Our method consists of deep Long-Short Term Memory Recurrent Neural Networks (LSTM-RNN) for dynamic Arousal and Valence regression, using acoustic and psychoacoustic features extracted from the songs that have been previously proven as effective for emotion prediction in music. Results on the challenge test demonstrate an excellent performance for Arousal estimation (r = 0.613 ± 0.278), but not for Valence (r = 0.026 ± 0.500). Issues regarding the quality of the test set annotations' reliability and distributions are indicated as plausible justifications for these results. By using a subset of the development set that was left out for performance estimation, we could determine that the performance of our approach may be underestimated for Valence (Arousal: r = 0.596 ± 0.386; Valence: r = 0.458 ± 0.551).","bibtype":"inproceedings","author":"Coutinho, Eduardo and Trigeorgis, George and Zafeiriou, Stefanos and Schuller, Björn","editor":"Larson, Martha and Ionescu, Bogdan and Sjöberg, Mats and Anguera, Xavier and Poignant, Johann and Riegler, Michael and Eskevich, Maria and Hauff, Claudia and Sutcliffe, Richard and Jones, Gareth J.F. and Yang, Yi-Hsuan and Soleymani, Mohammad and Papadopoulos, Symeon","booktitle":"CEUR Workshop Proceedings","keywords":"article,conference","bibtex":"@inproceedings{\n title = {Automatically estimating emotion in music with deep long-short term memory recurrent neural networks},\n type = {inproceedings},\n year = {2015},\n pages = {1-3},\n volume = {1436},\n websites = {https://www.scopus.com/inward/record.uri?eid=2-s2.0-84989923198&partnerID=40&md5=f0d407c8ffb96c19f299c78435558371,http://ceur-ws.org/Vol-1436/Paper64.pdf},\n month = {9},\n publisher = {CEUR},\n city = {Wurzen, Germany},\n id = {d43d00f9-3521-3815-bd81-243f874a33a5},\n created = {2024-08-09T12:19:57.621Z},\n file_attached = {true},\n profile_id = {ffa9027c-806a-3827-93a1-02c42eb146a1},\n group_id = {da2a8249-fdf4-3036-ba56-7358198a1600},\n last_modified = {2024-08-09T12:20:58.011Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n citation_key = {coutinho2015automaticallynetworks},\n source_type = {inproceedings},\n notes = {<b>From Duplicate 1 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, Eduardo; Trigeorgis, George; Zafeiriou, Stefanos; Schuller, Björn)<br/></b><br/><b>From Duplicate 1 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, Eduardo; Trigeorgis, George; Zafeiriou, Stefanos; Schuller, Björn)<br/></b><br/>cited By 7<br/><br/><b>From Duplicate 2 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, Eduardo; Trigeorgis, George; Zafeiriou, Stefanos; Schuller, Björn)<br/></b><br/><b>From Duplicate 1 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, E; Trigeorgis, G; Zafeiriou, S; Schuller, B)<br/></b><br/>cited By 7<br/><br/><b>From Duplicate 2 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, Eduardo; Trigeorgis, George; Zafeiriou, Stefanos; Schuller, Björn)<br/></b><br/><b>From Duplicate 1 (<i>Automatically estimating emotion in music with deep long-short term memory recurrent neural networks</i> - Coutinho, E; Trigeorgis, G; Zafeiriou, S; Schuller, B)<br/></b><br/>cited By 7},\n private_publication = {false},\n abstract = {In this paper we describe our approach for the MediaEval's \"Emotion in Music\" task. Our method consists of deep Long-Short Term Memory Recurrent Neural Networks (LSTM-RNN) for dynamic Arousal and Valence regression, using acoustic and psychoacoustic features extracted from the songs that have been previously proven as effective for emotion prediction in music. Results on the challenge test demonstrate an excellent performance for Arousal estimation (r = 0.613 ± 0.278), but not for Valence (r = 0.026 ± 0.500). Issues regarding the quality of the test set annotations' reliability and distributions are indicated as plausible justifications for these results. By using a subset of the development set that was left out for performance estimation, we could determine that the performance of our approach may be underestimated for Valence (Arousal: r = 0.596 ± 0.386; Valence: r = 0.458 ± 0.551).},\n bibtype = {inproceedings},\n author = {Coutinho, Eduardo and Trigeorgis, George and Zafeiriou, Stefanos and Schuller, Björn},\n editor = {Larson, Martha and Ionescu, Bogdan and Sjöberg, Mats and Anguera, Xavier and Poignant, Johann and Riegler, Michael and Eskevich, Maria and Hauff, Claudia and Sutcliffe, Richard and Jones, Gareth J.F. and Yang, Yi-Hsuan and Soleymani, Mohammad and Papadopoulos, Symeon},\n booktitle = {CEUR Workshop Proceedings},\n keywords = {article,conference}\n}","author_short":["Coutinho, E.","Trigeorgis, G.","Zafeiriou, S.","Schuller, B."],"editor_short":["Larson, M.","Ionescu, B.","Sjöberg, M.","Anguera, X.","Poignant, J.","Riegler, M.","Eskevich, M.","Hauff, C.","Sutcliffe, R.","Jones, G., J.","Yang, Y.","Soleymani, M.","Papadopoulos, S."],"urls":{"Paper":"https://bibbase.org/service/mendeley/ffa9027c-806a-3827-93a1-02c42eb146a1/file/7e37945f-c53d-4cfb-5be3-d5ece5bd879f/2015___Coutinho_et_al___Automatically_estimating_emotion_in_music_with_deep_long_short_term_memory_recurrent_neural_netw.pdf.pdf","Website":"https://www.scopus.com/inward/record.uri?eid=2-s2.0-84989923198&partnerID=40&md5=f0d407c8ffb96c19f299c78435558371,http://ceur-ws.org/Vol-1436/Paper64.pdf"},"biburl":"https://bibbase.org/service/mendeley/ffa9027c-806a-3827-93a1-02c42eb146a1","bibbaseid":"coutinho-trigeorgis-zafeiriou-schuller-automaticallyestimatingemotioninmusicwithdeeplongshorttermmemoryrecurrentneuralnetworks-2015","role":"author","keyword":["article","conference"],"metadata":{"authorlinks":{"coutinho, e":"http://138.253.69.159:8000/publications.html"}},"downloads":1},"search_terms":["automatically","estimating","emotion","music","deep","long","short","term","memory","recurrent","neural","networks","coutinho","trigeorgis","zafeiriou","schuller"],"keywords":["article","conference"],"authorIDs":["58e62eafe417243f58000008","58e63ebe8f93a1ae58000053","59b68df3dcf5fb104c000037","5c0e7cf57d11e01000000150","6Z276FjCJdxu2qttA","7HsugpwDMrLzay5kC","FShD3dRw9yuvcgPCs","HmyGQqdmpxNriPHXS","LxZfJ4SZafxuJzeKG","S2CzJg64apkJsPsHA","gSef4eEL8c5wuNF7w","hzjyyz7CdSRZyRptY","mo4CFXJ7ukAMT9nho","nxyiocLrG567rGYAm","oenDqaFMmkBMZtjuK","tE6ysWmRvSH5uoPFW","tvZgz2JnB9BN8znPF"],"dataSources":["Tcd3cXtdQsiKHPZsW","YqW8pMoihb7JazZcx","ya2CyA73rpZseyrZ8","HEzruaK9ZuLGaJrsP","jDf6YRbdcAMnNSFh6","2252seNhipfTmjEBQ"]}