Data Scientist | AI Lab at 3M
Ph.D in Computer Science
(interactive) Machine Learning | Audio Signal Processing | HCI
My research interests inlucde machine learning, audio signal processing (e.g., sound event recognition), intelligent interactive system, multimedia information retrieval, and human-in-the-loop interface. I also enjoy working on a musical interface and interactive media art.
3/30/2021: I was invited to Conversations on Applied AI Podcast and talked about my research on AI and audio. The episode is up now: link
04/28/2020: I’ve succefully defended my PhD disseration (Title: Sound Event Annotation and Detection with Less Human Effort)
04/14/2020: I’ve released an audio embedding model, M-VGGish which was used in my recent works: Link
9/4/2019: I gave a talk about my work, “Self-supervised Attention Model for Weakly Labeled Audio Event Classification” at EUSIPCO 2019 in A Coruña, Spain.
8/29/2019: My co-authored journal paper, “Learning to Build Natural Audio Production Interfaces” got published in Arts
8/24/2019: My co-authored paper, “Classifying non-speech vocals: Deep vs Signal Processing Representations” has been accepted from DCASE 2019
7/15/2019: My paper, “Sound Event Detection Using Point-labeled Data” has been accepted from WASPAA 2019
7/01/2019: My DCASE submission (task 5) got 3rd place out of 22 systems competing. (2nd in team rankings). Read more about the challenge and the results: click.
6/27/2019: I am giving a talk about “A Human-in-the-loop system for labeling sound events in audio recordings” at Midwest Music and Audio Day 2019 (MMAD) in Indiana University, Bloomington, USA.
6/03/2019: My paper, “Self-supervised Attention Model for Weakly Labeled Audio Event Classification” has been accepted from EUSIPCO 2019
5/11/2019: I am presenting my work, “Improving Content-based Audio Retrieval by Vocal Imitation Feedback” at ICASSP 2019 in Brighten, UK.
3/16/2019: I am giving a talk about my work, “A Human-in-the-loop System for Sound Event Detection and Annotation” at IUI 2019 in Los Angeles, USA.
11/19/2018: My sound classification model got 3rd place (out of 23 systems competing) in Making Sense of Sounds Data Challenge, 2018
11/19/2018: I am presenting my work, “Vocal Imitation Set: a dataset of vocally imitated sound events using the AudioSet ontology” at DCASE 2018 in Surrey, UK.