Skip to main content
menu

Digital Music Research Network

DMRN+16: Digital Music Research Network One-day Workshop 2021

 

Queen Mary University of London

Tuesday 21st December 2021

 

News

 

Keynote speakers

We have confirmed two keynote speakers.

  • Prof Sophie Scott - (UCL)

Title: "Sound on the brain - insights from functional neuroimaging and neuroanatomy"

  • Prof Gus Xia - (NYU Shanghai)

Title: "Learning interpretable music representations: from human stupidity to artificial intelligence"


 DMRN+16 is sponsored by

The UKRI Centre for Doctoral Training in Artificial Intelligence and Music (AIM); a leading PhD research programme aimed at the Music/Audio Technology and Creative Industries, based at Queen Mary University of London.

A new AIM CDT call for PhD positions will open in December 2021. There will be at least 12 fully funded PhD positions on AI + music topics.

 

Keynote talks

 

Keynote 1. Prof. Sophie Scott -Director, Institute of Cognitive Neuroscience, UCL.

Title: "Sound on the brain - insights from functional neuroimaging and neuroanatomy"

Abstract

In this talk I will use functional imaging and models of primate neuroanatomy to explore how sound is processed in the human brain. I will demonstrate that sound is represented cortically in different parallel streams. I will expand this to show how this can impact on the concept of auditory perception, which arguably incorporates multiple kinds of distinct perceptual processes.  I will address the roles that subcortical processes play in this, and also the contributions from hemispheric asymmetries.


 Keynote 2: Prof. Gus Xia - Assistant Professor at NYU Shanghai

Title: "Learning interpretable music representations: from human stupidity to artificial intelligence"

Abstract

Gus has been leading the Music X Lab in developing intelligent systems that help people better compose and learn music. In this talk, he will show us the importance of music representation for both humans and machines, and how to learn better music representations via the design of inductive bias. Once we got interpretable music representations, the potential applications are limitless.

 

Call for Contributions

The Digital Music Research Network (DMRN) aims to promote research in the area of digital music, by bringing together researchers from UK and overseas universities, as well as industry, for its annual workshop. The workshop will include invited and contributed talks and posters. The workshop will be an ideal opportunity for networking with other people working in the area. 

 

* Call for Contributions

You are invited to submit a proposal for a "talk" and/or a "poster" to be presented at this event.

TALKS may range from the latest research, through research overviews or surveys, to opinion pieces or position statements, particularly those likely to be of interest to an interdisciplinary audience. Due to the online format, we plan to keep talks to about 10 minutes each, depending on the number of submissions. Short announcements about other items of interest (e.g. future events or other networks) are also welcome.

POSTERS can be on any research topic of interest to the members of the network. Posters will be displayed virtually via the online platform gather.town.

The abstracts of presentations will be collated into a digest and distributed on the day.

 

* Submission

Please prepare your talk or poster proposal in the form of an abstract (1 page A4, using the template PDF-DMRN+16-Template [PDF 155KB], Word-DMRN+16-Template [DOC 94KB], One-page Template LaTex [403KB]). Submit it via email to dmrn@lists.eecs.qmul.ac.uk giving the following information about your presentation:

  • Authors
  • Title
  • Abstract
  • Preference for talk or poster (or "no preference").

 

* Deadlines

Extended to Friday 26th November - Abstract submission deadline

29 Nov 2021: Notification of acceptance

17 Dec 2021: Registration deadline

21 Dec 2021: DMRN+16 Workshop

 

Registration

The event will be online, registration is mandatory.

Register here

Ths information with the details and the links to join the workshop will be send closer to the workshop.

 

 

 Programme

 

The final programme will be published on Monday 6th December. The provisional programme is:

 

10:00

Welcome, Simon Dixon (Queen Mary University of London)

10:10

KEYNOTE

 "Sound on the brain - insights from functional neuroimaging and neuroanatomy", Prof Sophie Scott - (Institute of Cognitive Neuroscience - UCL)

 

10:50

Building style-aware neural MIDI synthesizers using simplified differentiable DSP approach, Sergey Grechin and Ryan Groves (Infinite Album)

11:00

Completing Audio Drum Loops with Transformer Neural Networks, Teresa Pelinski (Queen Mary University of London), Behzad Haki and Sergi Jordà (Pompeu Fabra University)

11:10

Evaluation of GPT-2-based Symbolic Music Generation, Berker Banar and Simon Colton (Queen Mary University of London)

11:20

NASH: the Neural Audio Synthesis Hackathon, Ben Hayes, Cyrus Vahidi and Charalampos Saitis (Queen Mary University of London)

11:30

Designing a synthesiser to elicit a feeling of perceived tension, Connor Welham, Bruno Fazenda, and Duncan Williams (University of Salford)

11:40

Is Automatically Transcribed Data Reliable Enough for Expressive Piano Performance Research?, Huan Zhang and Simon Dixon (Queen Mary University of London)

11:50

CAMAT: Computer Assisted Music Analysis Toolkit, Egor Poliakov (HMT Leipzig) and Christon R. Nadar (Fraunhofer IDMT)

 12:00

Lunch break

 

13:15

KEYNOTE

 "Learning interpretable music representations: from human stupidity to artificial intelligence". Assistant Prof Gus Xia - (NYU Shanghai)

 13:55

Announcements and Intro to Gather Town

14:00

POSTER SESSION

 Open poster session where the participant will be able to view the poster and chat with the authors.

16:00

Close

 

 

Poster session

 

1

Sketching Sounds: Using sound-shape associations to build a sketch- based sound synthesizer, Sebastian Löbbers and George Fazekas (Queen Mary University of London)

2

Everyday Sound Recognition with Limited Annotations, Jinhua Liang, Huy Phan and Emmanouil Benetos (Queen Mary University of London)

 

3

Generating Comments from Music and Lyrics, Yixiao Zhang and Simon Dixon (Queen Mary University of London)

4

AI-Assisted FM Synthesis, Franco Caspe, Mark Sandler and Andrew McPherson (Queen Mary University of London)

5

Algorithmic Music Composition for The Environment, Rosa Park (San Francisco State University)

6

The Vienna Philharmonic's New Year's Concert Series: A Corpus for Digital Musicology and Performance Science, David M. Weigl and Werner Goebl (University of Music and Performing Arts Vienna)

7

An Interactive Tool for Visualising Musical Performance Subtleties, Yucong Jiang (University of Richmond)

8

A Benchmark Dataset to Study Microphone Mismatch Conditions for Piano Multipitch Estimation on Mobile Devices, Jakob Abeßer, Franca Bittner, Maike Richter, Marcel Gonzalez and Hanna Lukashevich (Fraunhofer IDMT)

9

Looking at the Future of Data-Driven Procedural Audio, Adrián Barahona-Ríos (University of York)

10

Making graphical scores accessible to visually impaired people: A haptic interactive installation, Christina Karpodini

11

Acoustic Representations for Perceptual Timbre Similarity, Cyrus Vahidi, Ben Hayes, Charalampos Saitis and George Fazekas (Queen Mary University of London)

12

Investigating a computational methodology for quantitive analysis of singing performance style, Yukun Li, Polina Proutskova, Zhaoxin Yu and Simon Dixon (Queen Mary University of London)

13

Variational Auto Encoding and Cycle-Consistent Adversarial Networks for Timbre Transfer, Russell Sammut Bonnici, Martin Benning and Charalampos Saitis (Queen Mary University of London)

 

 

 

Contact information

Alvaro Bort (a.bort@qmul.ac.uk)

EECS, Queen Mary University of London
Mile End Road, London E1 4NS, UK.

Return to top