The Creative Audio Synthesis and Interfaces Workshop

On 15 July, 2025, The Creative Audio Synthesis and Interfaces Workshop was held at Queen Mary University of London, organised by AIM COMMA Lab members Jordie Shier, Haokun Tian, and Charalampos Saitis, and supported by the UKRI Centre for Doctoral Training in Artificial Intelligence and Music (AIM) at the Centre for Digital Music (C4DM).

This one-day workshop included a series of talks exploring the intersection of creative audio synthesis and AI-enabled synthesizer programming.
Topics included evolutionary algorithms for sound exploration, synthesizer sound matching, timbre transfer, timbre-based control, reinforcement learning, differentiable digital signal processing, representation learning, and human-machine co-creativity.

Earlier this year, the AIM CDT visited the RITMO Centre for Interdisciplinary Studies in Rhythm, Time, and Motion at the University of Oslo in Norway.
A shared interest in creative applications of audio synthesis and novel interface designs was established during this visit, motivating this follow-up workshop at QMUL.
Researchers from RITMO, The Open University, and the AIM CDT at QMUL were invited to
share their work, engage in critical discussion, and map directions for
future work. See below for a summary of talks with links to presentation recordings.

An evening concert showcased musical applications of technical implementations discussed during the workshop, grounding these discussions in real-world artistic contexts.

Invited Talks

Designing Percussive Timbre Remappings: Negotiating Audio Representations and Evolving Parameter Spaces

Facilitating serendipitous sound discoveries with simulations of open-ended evolution

Autonomous control of synthesis parameters with listening-based reinforcement learning

Can a Sound Matching Model Produce Audio Embeddings that Align with Timbre Similarity Rated by Humans?

GuitarFlow: Realistic Electric Guitar Synthesis From Tablatures via Flow Matching and Style Transfer

  • Jackson Loth — Queen Mary University of London

Timbre latent space transformations for interactive musical systems oriented to timbral music-making

Why Synthesizer Parameter Estimation Is Hard and How to Make it Easy

Perceptually Aligned Deep Image Sonification

Modulation Discovery with Differentiable Digital Signal Processing

Musical Performances and Demos

Experience Replay (Performance)

  • Vincenzo Madaghiele — University of Oslo

Weaving (Performance)

  • Balint Laczko — University of Oslo

Phylogeny (Demo)

  • Björn Thor Jónsson — University of Oslo

AIM/C4DM team wins Query-by-Vocal Imitation Challenge

Congratulations to AIM members Aditya Bhattacharjee and Christos Plachouras, and C4DM member Sungkyun Chang who secured first place at the Query-by-Vocal Imitation (QbVI) Challenge, held as part of the AES International Conference on Artificial Intelligence and Machine Learning for Audio conference (AES AIMLA 2025) taking place from September 8-10, 2025.

The winning entry addressed the task, which entails retrieving relevant audio clips from a database using only a vocal imitation as a query. This is a particularly complex problem due to the variability in how people vocalise sounds and the acoustic diversity across sound categories. Successful approaches must bridge the gap between vocal and non-vocal audio, while handling the unpredictability of human-generated imitations.

The team’s submission, titled “Effective Finetuning Methods for Query-by Vocal Imitation”, advances the state-of-the-art in QbVI by integrating a triplet-based regularisation objective with supervised contrastive learning. This method addresses the issue of limited data by sampling from an unused subset of the VocalSketch dataset, which comprises practice recordings and human-rejected vocal imitations. While this data may not be suitable for positive matches, the vocal imitation data is useful for creating confounding examples during training. More specifically, this increases the size of the pool of negative examples, which is utilised by the added regularisation method.

The proposed method surpassed state-of-the-art methods for both subjective and objective evaluation metrics, opening up scope for product-based innovations and software tools that can be used by artists to effectively search large repositories of sound effects.


AIM at ISMIR 2025

ISMIR 2025 logoOn 21-25 September 2025, several AIM researchers will participate at the 26th International Society for Music Information Retrieval Conference (ISMIR 2025). ISMIR is the leading conference in the field of music informatics, and is currently the top-cited publication for Music & Musicology (source: Google Scholar). This year ISMIR will take place onsite in Daejeon, Korea.

Similar to previous years, AIM will have a strong presence at ISMIR 2025.

In the Scientific Programme, the following papers are authored/co-authored by AIM members:

The following Tutorials will be co-presented by AIM PhD students Rodrigo Diaz and Julien Guinot:

  • Differentiable Physical Modeling Sound Synthesis: Theory, Musical Application, and Programming (Jin Woo Lee, Stefan Bilbao, Rodrigo Diaz)
  • Self-supervised Learning for Music – An Overview and New Horizons (Julien Guinot, Alain Riou, Yuexuan Kong, Marco Pasini, Gabriel Meseguer-Brocal, Stefan Lattner)

The following journal papers published at TISMIR which are co-authored by AIM members will be presented at the conference:

As part of the MIREX public evaluations:

Finally, on the organisational side:

See you at Daejeon!