Clara is working on augmented reality approaches to improve sound comfort in indoor environments, specifically in shared workspaces. She is a PhD student, supervised by Nicolas Misdariis from the Ircam in Paris and Mathieu Lagrange from the SIMS team at LS2N.
Tag: in English
This post is available in English.
Towards better visualizations of urban sound environments: insights from interviews @ INTERNOISE 2024
Urban noise maps and noise visualizations traditionally provide macroscopic representations of noise levels across cities. However, those representations fail at accurately gauging the sound perception associated with these sound environments, as perception highly depends on the sound sources involved. This paper aims at analyzing the need for the representations of sound sources, by identifying the urban stakeholders for whom such representations are assumed to be of importance. Through spoken interviews with various urban stakeholders, we have gained insight into current practices, the strengths and weaknesses of existing tools and the relevance of incorporating sound sources into existing urban sound environment representations. Three distinct use of sound source representations emerged in this study: 1) noise-related complaints for industrials and specialized citizens, 2) soundscape quality assessment for citizens, and 3) guidance for urban planners. Findings also reveal diverse perspectives for the use of visualizations, which should use indicators adapted to the target audience, and enable data accessibility.
“Sensing the City Using Sound Sources: Outcomes of the CENSE Project” @ Urban Sound Symposium
Full program at: https://urban-sound-symposium.org/program/
Challenge on Sound Scene Synthesis: Evaluating Text-to-Audio Generation @ NeurIPS Audio Imagination workshop
Despite significant advancements in neural text-to-audio generation, challenges persist in controllability and evaluation. This paper addresses these issues through the Sound Scene Synthesis challenge held as part of the Detection and Classification of Acoustic Scenes and Events 2024. We present an evaluation protocol combining objective metric, namely Fréchet Audio Distance, with perceptual assessments, utilizing a structured prompt format to enable diverse captions and effective evaluation. Our analysis reveals varying performance across sound categories and model architectures, with larger models generally excelling but innovative lightweight approaches also showing promise. The strong correlation between objective metrics and human ratings validates our evaluation approach. We discuss outcomes in terms of audio quality, controllability, and architectural considerations for text-to-audio synthesizers, providing direction for future research.
BirdVox in MIT Technology Review
Vincent speaks to MIT Technology Review on the past, present, and future of machine learning for bird migration monitoring.
“Advancements in Bird Communication Studies” abstract deadline
Please submit your abstract (max. 200 words) towards a special session at Forum Acusticum / Euronoise, to be held in Málaga (Spain) on June 23–26, 2025.
Can machines learn filterbank design?
A talk by Vincent Lostanlen at the Acoustics Research Institute (ARI) of the Austrian Academy of Sciences (ÖAW) in Vienna.
Phantasmagoria @ Academy of Fine Arts of Munich
Invited talk: Yuexuan Kong
“STONE: Self-supervised tonality estimator”, at École Centrale de Nantes, amphi E, 2pm.
Introducing: Reyhaneh Abbasi
Reyhaneh is working on the generation of mouse ultrasonic vocalizations, with applications to animal behavior research.