Abigail Leon has successfully defended her PhD today with main supervisor Prof. Yuen Chau and myself as co-supervisor. Abigail has successfully explored how deep learning techniques can be used to denoise en demodulate complex underwater acoustic communication signals, and has performed some sea-trials to gather data for this. Since the PhD is under an NDA with Thales, we cannot post it, however, check out some of Abigail's papers here (more to come after the review process ends).
- Research assistant / postdoc jobs in Music/Audio and AI
- New article in Sensors: Deep Neural Network-Based Respiratory Pathology Classification Using Cough Sounds
- aiSTROM -- A roadmap for developing a successful AI strategy
- Book chapter on Musical stylometry: Characterisation of music
- Joint internship with Sounders Music
- Meet My Lab - podcast from Euraxess
- New roadmap paper on the role of music technology for health care and well-being
- Three IJCNN papers from the AMAAI lab this year!
- Keynote at DMRN on controllable music generation
- New paper on Underwater Acoustic Communication Receiver Using Deep Belief Network
AI.SG is offering scholarships to top students who are interested in AI research. My lab at Singapore University of Technology and Design has positions for those interested in research on AI for audio/music/finance and NLP/emotion. More information about the lab and our projects: dorienherremans.com
Since this is an elite scholarship you can only apply if you have:
- high GPA
- Master's degree, or existing core ranked publications
Looking for a tool to extract spectrograms on the fly, integrated as a layer in PyTorch? Look no further than nnAudio, a toolbox developed by PhD student Raven (Cheuk Kin Wai): https://github.com/KinWaiCheuk/nnAudio
nnAudio is available in pip (pip install nnaudio), full documentation available on the github page. Also check out our dedicated paper:
In a collaboration with IIT, India and SUTD, we've published a paper on our new perceptionGAN system in the Proceedings of the 4th Int. Conf. on Imaging, Vision and Pattern Recognition (IVPR), and 9th Int. Conf. on Informatics, Electronics & Vision (ICIEV). Read the preprint.
SUTD's AMAAI lab is organizing online seminars of (worldwide) graduate students active in Music Information Retrieval (MIR), or more generally music/audio and AI. The aim is to connect different labs working on similar topics and enable international collaboration. Participating universities include SUTD, QMUL,...
The Webinars will be organized on Wednesdays at 4pm Singapore time (9am UK time - 10am EU time).
Interesting is doing MIR research in Singapore? Let me know, EU has released the call for Marie-Curie fellowships. Having been a MSCA fellow myself I can highly recommend these! https://ec.europa.eu/info/funding-tenders/opportunities/portal/screen/op...
Over the last few years, there’s been a steady growth in revenue from digital music. In just six years, revenue from music streaming moved from zero to 40 percent of the overall global recorded music industry revenues, according to a report by IFPI. With revenues to the tune of 11.2 billion dollars a year, the digital model is only set to grow. So “is there still room for a traditional record company?”
Read more in the interview I gave to the SUTD Aspire newsletter about my recent keynote in TechHR.
The special issue on Deep Learning for Music and Audio has been published in Neural Computing and Applications (Impact factor 4.664 (2018)). The guest editors for this special issue were Prof. Ching-Hua Chuan from the University of Miami and myself.
PhD student Yin-Jyun Luo got his paper on 'Singing voice conversion with disentangled representations of singer and vocal technique using variational autoencoders' accepted for the upcoming ICASSP conference in Barcelona Spain. You can read the preprint on Arxiv.
Our team at SUTD Game Lab, directed by Prof. Dorien Herremans, is looking for:
PhD students in Game Research with focus on AI or AR
Do you love gaming and want to make it your specialisation? We are a vibrant team at Singapore University of Technology and Design that create serious games for industry and academia. Our team consists of artists, game designers, and game developers. As a PhD student, you will work on your own research ideas, possibly combined with one of the team's projects, on topics such as:
Phd student Thao Phuang's paper on "Multimodal Deep Models for Predicting Affective Responses Evoked by Movies" was awarded best student paper at the 2nd International Workshop on Computer Vision for Physiological Measurement as part of ICCV in Seoul, South Korea. The paper explores how models based on video and audio can predict emotion of movies:
Together with my PhD student Thao and Prof. Gemma Roig (MIT/Frankfurt University), a new paper was published on "Multimodal Deep Models for Predicting Affective Responses Evoked by Movies" in the Proceedings of the 2nd International Workshop on Computer Vision for Physiological Measurement as part of ICCV. Seoul, South Korea. 2019. A preprint is available here.
Together with Dr. Kat Agres (NUS, Singapore) and Prof. Louis Bigo (University of Lille, France), I recently explored how harmonic structure influences altered states in uplifting trance music. "The Impact of Musical Structure on Enjoyment and Absorptive Listening States in Trance Music" is available as a chapter in Music and Consciousness II, a book published by Ruth Herbert, Eric Clarke and David Clarke.
One of the most comprehensible machine learning instructors (Andrew Ng), has just released a new book. I found it an interesting and smooth read with focus on a few important issues that any data science / AI person will encounter. Highly recommend the read. More info on the book and how to get it.
Just published a new article with my PhD student Thao Ha Thi Phuong and Prof. Gemma Roig on 'Multimodal Deep Models for Predicting Affective Responses Evoked by Movies'. The paper will be published in the proceedings of the 2nd International Workshop on Computer Vision for Physiological Measurement as part of ICCV; and will be presented by Thao in Seoul, South Korea. Anybody interested can download the preprint article here (link coming soon!). The source code of our model is available on github.