Design Recommendations for a Collaborative Game of Bird Call Recognition Based on Internet of Sound Practices

Detailed info Design Recommendations for a Collaborative Game of Bird Call Recognition Based on Internet of Sound Practices Authors:Rovithis, Emmanouel; Moustakas, Nikolaos; Vogklis, Konstantinos; Drossos, Konstantinos; Floros, AndreasTitle:Design Recommendations for…

Continue ReadingDesign Recommendations for a Collaborative Game of Bird Call Recognition Based on Internet of Sound Practices

Enriched Music Representations With Multiple Cross-Modal Contrastive Learning

Modeling various aspects that make a music piece unique is a challenging task, requiring the combination of multiple sources of information. Deep learning is commonly used to obtain representations using various sources of information, such as the audio, interactions between users and songs, or associated genre metadata.

Continue ReadingEnriched Music Representations With Multiple Cross-Modal Contrastive Learning

Continual Learning for Automated Audio Captioning Using The Learning Without Forgetting Approach

  • Post category:Publications

Automated audio captioning (AAC) is the task of automatically creating textual descriptions (i.e. captions) for the contents of a general audio signal. Most AAC methods are using existing datasets to optimize and/or evaluate upon.

Continue ReadingContinual Learning for Automated Audio Captioning Using The Learning Without Forgetting Approach

Fairness and underspecification in acoustic scene classification: The case for disaggregated evaluations

Underspecification and fairness in machine learning (ML) applications have recently become two prominent issues in the ML community. Acoustic scene classification (ASC) applications have so far remained unaffected by this discussion, but are now becoming increasingly used in real-world systems where fairness and reliability are critical aspects.

Continue ReadingFairness and underspecification in acoustic scene classification: The case for disaggregated evaluations

Automatic Analysis of the Emotional Content of Speech in Daylong Child-Centered Recordings from a Neonatal Intensive Care Unit

  • Post category:Publications

Researchers have recently started to study how the emotional speech heard by young infants can affect their developmental outcomes. As a part of this research, hundreds of hours of daylong recordings from preterm infants’ audio environments were collected from two hospitals in Finland and Estonia in the context of so-called APPLE study.

Continue ReadingAutomatic Analysis of the Emotional Content of Speech in Daylong Child-Centered Recordings from a Neonatal Intensive Care Unit