AIM CDT Cohorts |
---|
First Cohort - 2019-2024 |
Second Cohort - 2020-2025 |
Third Cohort - 2021-2026 |
Fourth Cohort - 2022-2027 |
Fifth Cohort - 2023-2028 |
Sixth Cohort - 2024-2029 |
First AIM cohort students (2019-2024):
PhD Student | Project title |
---|---|
Adan Benito | Beyond the fret: gesture analysis on fretted instruments and its applications to instrument augmentation |
Berker Banar | Towards Composing Contemporary Classical Music using Generative Deep Learning |
Marco Comunità | Machine learning applied to sound synthesis models |
David Foster | Modelling the Creative Process of Jazz Improvisation |
Lele Liu | Automatic music transcription with end-to-end deep neural networks |
Ilaria Manco | Deep learning and multi-modal models for the music industry in collaboration with Universal Music Group |
Andrea Martelloni | Real-Time Gesture Classification on an Augmented Acoustic Guitar using Deep Learning to Improve Extended-Range and Percussive Solo Playing |
Mary Pilataki-Manika | Polyphonic Music Transcription using Deep Learning in collaboration with Apple |
Saurjya Sarkar | New perspectives in instrument-based audio source separation |
Pedro Sarmento | Guitar-Oriented Neural Music Generation in Symbolic Format in collaboration with Holonic Systems Oy. |
Elona Shatri | Optical music recognition using deep learning in collaboration with Steinberg Media Technologies GmbH |
Cyrus Vahidi | Perceptual end to end learning for music understanding in collaboration with MUSIC Tribe Brands UK Limited |
Second AIM cohort students (2020-2025):
PhD Student | Project title |
---|---|
Corey Ford | Artificial Intelligence for Supporting Musical Creativity and Engagement in Child-Computer Interaction |
Max Graf | AI-Based Musical Co-Creation in Extended Realities: PERFORM-AI |
Madeline Hamilton | Perception and analysis of 20th and 21st century popular music |
Benjamin Hayes | Perceptually motivated deep learning approaches to creative sound synthesis |
Jiawen Huang | Lyrics Alignment For Polyphonic Music |
Harnick Khera | Informed source separation for multi-mic production in collaboration with BBC |
Yin-Jyun Luo | Industry-scale Machine Listening for Music and Audio Data in collaboration with Spotify |
Luca Marinelli | Gender-coded sound: A multimodal data-driven analysis of gender encoding strategies in sound and music for advertising |
Xavier Riley | Digging Deeper - expanding the “Dig That Lick” corpus with new sources and techniques |
Eleanor Row | Automatic micro-composition for professional/novice composers using generative models as creativity support tools |
Shubhr Singh | Audio Applications of Novel Mathematical Methods in Deep Learning |
Christian Steinmetz | Deep learning for high-fidelity audio and music production |
Jingjing Tang | End-to-End System Design for Music Style Transfer with Neural Networks |
Lewis Wolstanholme | Real-time instrument transformation and augmentation with deep learning |
Yixiao Zhang | Machine Learning Methods for Artificial Musicality in collaboration with Apple |
Third AIM cohort students (2021-2026):
PhD Student | Project title |
---|---|
Katarzyna Adamska | Predicting hit songs: multimodal and data-driven approach |
Sara Cardinale | Character-based adaptive generative music for film and video games using Deep Learning and Hidden Markov Models |
Franco Caspe | AI-assisted FM synthesis for sound design and control mapping |
Ruby Crocker | Time-based mood recognition in film music |
Carlos De La Vega Martin | Neural Drum Synthesis |
Bleiz MacSen Del Sette | The Sound of Care: researching the use of deep learning and sonification for the daily support of people with Chronic Pain |
Rodrigo Mauricio Diaz Fernandez | Hybrid Neural Methods for Sound Synthesis |
Andrew Edwards | Computational Models for Jazz Piano: Transcription, Analysis, and Generative Modeling |
Oluremi Samuel Oladotun Falawo | Embodiment in Intelligent Musical Systems |
Mariam Fayaz Torshizi | Music mood modelling using Knowledge graphs and Graph Neural Nets |
Yazhou Li | Virtual Placement of Objects in Acoustic Scenes |
Jackson Loth | Time to vibe together: cloud-based guitar and intelligent agent in collaboration with Hyvibe |
Teresa Pelinski Ramos | Sensor mesh as performance interface in collaboration with Bela |
Soumya Sai Vanka | Smart Channel strip using Neural audio processing in collaboration with Steinberg |
Chris Winnard | Music Interestingness in the Brain |
Xiaowan Yi | Composition-aware music recommendation system for music production in collaboration with Focusrite |
Huan Zhang | Computational Modeling of Expressive Piano Performance |
Fourth AIM cohort students (2022-2027):
PhD Student | Project title |
---|---|
James Bolt | Intelligent audio and music editing with deep learning |
Carey Bunks | Cover Song Identification in collaboration with Apple |
Adam Garrow | A computational model of music cognition using statistical learning of structures |
Ashley Noel-Hirst | Latent Spaces for Human-AI music generation |
Alexander Williams | User-driven deep music generation in digital audio workstations in collaboration with Sony |
Yinghao Ma | Self-supervision in machine listening in collaboration with Bytedance |
Jordan Shier | Real-time timbral mapping for synthesized percussive performance in collaboration with Ableton |
David Südholt | Machine learning of physical models for voice synthesis in collaboration with Nemisindo |
Tyler McIntosh | Expressive Performance Rendering for Music Generation Systems in collaboration with DAACI |
Christopher Mitcheltree | Deep Learning for Time-varying Audio and Parameter Modulations |
Ioannis Vasilakis | Active learning for interactive music transcription |
Chin-Yun Yu | Neural audio synthesis with expressiveness control |
Ningzhi Wang | Generative models for music audio representation and understanding in collaboration with Spotify |
Aditya Bhattacharjee | Self-supervision in Audio Fingerprinting |
Fifth AIM cohort students (2023-2028):
PhD Student | Project title |
---|---|
Bradley Aldous | Advancing music generation via accelerated deep learning |
Keshav Bhandari | Neuro-Symbolic Automated Music Composition |
Louis Bradshaw | Neuro-symbolic music models |
Julien Guinot | Beyond Supervised Learning for Musical Audio (in collaboration with Universal Music Group) |
Zixun (Nicolas) Guo | Towards Tonality-Aware Music Understanding: Modeling Complex Tonal Harmony |
Adam He | Neuro-evolved Heuristics for Meta-composition (in collaboration with DAACI) |
Gregor Meehan | Representation learning for musical audio using graph neural network-based recommender engines |
Marco Pasini | Fast and Controllable Music Generation (in collaboration with Sony SCL) |
Christos Plachouras | Deep learning for low-resource music |
Pablo Tablas De Paula | Machine Learning of Physical Models |
Haokun Tian | Timbre Tools for the Digital Instrument Maker |
Qing Wang | Multi-modal Learning for Music Understanding |
Yifan Xie | Film score composer AI assistant: generating expressive mockups (in collaboration with Spitfire Audio) |
Ece Yurdakul | Emotion-based Personalised Music Recommendation (in collaboration with Deezer) |
Farida Yusuf | Neural computing for auditory object analysis |
Qiaoxi Zhang | Multimodal AI for musical collaboration in immersive environments (in collaboration with PatchXR) |
Shuoyang Zheng | Explainability of AI Music Generation |
Xavi D'Cruz | Multimodal techniques for the control of procedural audio |
Sixth AIM cohort students (2024-2029):
PhD Student | Project title |
---|---|
Jinwen Zhou | Combining Deep Learning and Music Theory |
Minhui Lu | Applications of deep learning for improved synthesis of engine sounds |
Shangxuan Luo | Adaptive Music Generation for Video Game |
Yorgos Velissaridis | Smart EQ: Personalizing Audio with Context-aware AI using Listener Preferences and Psychological Factors |