Paper Digest: Recent Papers on AI for Music
Paper Digest Team extracted all recent AI for Music related papers on our radar, and generated highlight sentences for them. The results are then sorted by relevance & date. In addition to this ‘static’ page, we also provide a real-time version of this article, which has more coverage and is updated in real time to include the most recent updates on this topic.
Based in New York, Paper Digest is dedicated to helping people generate contents & reason over unstructured data. Different from black-box approaches, we build deep models on semantics, which allows results to be produced with explainations. Such models power this website, and are behind our services including “search engine”, “summarization”, “question answering”, and “literature review”.
If you do not want to miss interesting academic papers, you are welcome to sign up our daily paper digest service to get updates on new papers published in your area every day. You are also welcome to follow us on Twitter and Linkedin to get updated with new conference digests.
Paper Digest Team
New York City, New York, 10017
team@paperdigest.org
TABLE 1: Paper Digest: Recent Papers on AI for Music
Paper | Author(s) | Source | Date | |
---|---|---|---|---|
1 | Efficient Supervised Training of Audio Transformers for Music Representation Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we address music representation learning using convolution-free transformers. |
Pablo Alonso-Jiménez; Xavier Serra; Dmitry Bogdanov; | arxiv-cs.SD | 2023-09-28 |
2 | Predicting Performance Difficulty from Piano Sheet Music Images Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This architecture is adapted to our task by introducing an encoding scheme that reduces the encoded sequence length to one-eighth of the original size. In terms of evaluation, we consider five datasets — more than 7500 scores with up to 9 difficulty levels — , two of them particularly compiled for this work. |
Pedro Ramoneda; Jose J. Valero-Mas; Dasaem Jeong; Xavier Serra; | arxiv-cs.SD | 2023-09-28 |
3 | FineDance: A Fine-grained Choreography Dataset for 3D Full Body Dance Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Generating full-body and multi-genre dance sequences from given music is a challenging task, due to the limitations of existing datasets and the inherent complexity of the fine-grained hand motion and dance genres. To address these problems, we propose FineDance, which contains 14.6 hours of music-dance paired data, with fine-grained hand motions, fine-grained genres (22 dance genres), and accurate posture. |
RONGHUI LI et. al. | iccv | 2023-09-27 |
4 | Video Background Music Generation: Dataset, Method and Evaluation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This is a challenging task since it requires music-video datasets, efficient architectures for video-to-music generation, and reasonable metrics, none of which currently exist. To close this gap, we introduce a complete recipe including dataset, benchmark model, and evaluation metric for video background music generation. |
LE ZHUO et. al. | iccv | 2023-09-27 |
5 | TM2D: Bimodality Driven 3D Dance Generation Via Music-Text Integration Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a novel task for generating 3D dance movements that simultaneously incorporate both text and music modalities. |
KEHONG GONG et. al. | iccv | 2023-09-27 |
6 | Synthia’s Melody: A Benchmark Framework for Unsupervised Domain Adaptation in Audio Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We, in part, attribute this to the lack of an appropriate benchmark dataset. To address this gap, we present Synthia’s melody, a novel audio data generation framework capable of simulating an infinite variety of 4-second melodies with user-specified confounding structures characterised by musical keys, timbre, and loudness. |
Chia-Hsin Lin; Charles Jones; Björn W. Schuller; Harry Coppock; | arxiv-cs.SD | 2023-09-26 |
7 | CrossSinger: A Cross-Lingual Multi-Singer High-Fidelity Singing Voice Synthesizer Trained on Monolingual Singers Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose CrossSinger, which is a cross-lingual singing voice synthesizer based on Xiaoicesing2. |
Xintong Wang; Chang Zeng; Jun Chen; Chunhui Wang; | arxiv-cs.SD | 2023-09-22 |
8 | Passage Summarization with Recurrent Models for Audio-Sheet Music Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, two challenges that arise out of this strategy are the requirement of strongly aligned data to train the networks, and the inherent discrepancies of musical content between audio and sheet music snippets caused by local and global tempo differences. In this paper, we address these two shortcomings by designing a cross-modal recurrent network that learns joint embeddings that can summarize longer passages of corresponding audio and sheet music. |
Luis Carvalho; Gerhard Widmer; | arxiv-cs.SD | 2023-09-21 |
9 | Performance Conditioning for Diffusion-Based Multi-Instrument Music Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: As the main contribution of this work, we propose enhancing control of multi-instrument synthesis by conditioning a generative model on a specific performance and recording environment, thus allowing for better guidance of timbre and style. |
Ben Maman; Johannes Zeitler; Meinard Müller; Amit H. Bermano; | arxiv-cs.SD | 2023-09-21 |
10 | Self-Supervised Contrastive Learning for Robust Audio-Sheet Music Retrieval Systems Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, the scarcity of annotated data from real musical content affects the capability of such methods to generalize to real retrieval scenarios. In this work, we investigate whether we can mitigate this limitation with self-supervised contrastive learning, by exposing a network to a large amount of real music data as a pre-training step, by contrasting randomly augmented views of snippets of both modalities, namely audio and sheet images. |
Luis Carvalho; Tobias Washüttl; Gerhard Widmer; | arxiv-cs.SD | 2023-09-21 |
11 | Towards Robust and Truly Large-Scale Audio-Sheet Music Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this article we attempt to provide an insightful examination of the current developments on audio-sheet music retrieval via deep learning methods. |
Luis Carvalho; Gerhard Widmer; | arxiv-cs.SD | 2023-09-21 |
12 | K-pop Lyric Translation: Dataset, Analysis, and Neural-Modelling Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To broaden the scope of genres and languages in lyric translation studies, we introduce a novel singable lyric translation dataset, approximately 89\% of which consists of K-pop song lyrics. |
Haven Kim; Jongmin Jung; Dasaem Jeong; Juhan Nam; | arxiv-cs.CL | 2023-09-20 |
13 | Leveraging Negative Signals with Self-Attention for Sequential Music Recommendation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we investigate the use of transformer-based self-attentive architectures to learn implicit session-level information for sequential music recommendation. |
Pavan Seshadri; Peter Knees; | arxiv-cs.IR | 2023-09-20 |
14 | Investigating Personalization Methods in Text to Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we investigate the personalization of text-to-music diffusion models in a few-shot setting. |
Manos Plitsis; Theodoros Kouzelis; Georgios Paraskevopoulos; Vassilis Katsouros; Yannis Panagakis; | arxiv-cs.SD | 2023-09-20 |
15 | Popularity Degradation Bias in Local Music Recommendation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we study the effect of popularity degradation bias in the context of local music recommendations. |
April Trainor; Douglas Turnbull; | arxiv-cs.IR | 2023-09-20 |
16 | MelodyGLM: Multi-task Pre-training for Symbolic Melody Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose MelodyGLM, a multi-task pre-training framework for generating melodies with long-term structure. |
XINDA WU et. al. | arxiv-cs.SD | 2023-09-19 |
17 | Motif-Centric Representation Learning for Symbolic Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we aim to learn the implicit relationship between motifs and their variations via representation learning, using the Siamese network architecture and a pretraining and fine-tuning pipeline. |
Yuxuan Wu; Roger B. Dannenberg; Gus Xia; | arxiv-cs.SD | 2023-09-19 |
18 | HumTrans: A Novel Open-Source Dataset for Humming Melody Transcription and Beyond Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper introduces the HumTrans dataset, which is publicly available and primarily designed for humming melody transcription. |
Shansong Liu; Xu Li; Dian Li; Ying Shan; | arxiv-cs.SD | 2023-09-18 |
19 | Positive and Risky Message Assessment for Music Products Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a novel research problem: assessing positive and risky messages from music products. |
Yigeng Zhang; Mahsa Shafaei; Fabio Gonzalez; Thamar Solorio; | arxiv-cs.CL | 2023-09-18 |
20 | Unified Pretraining Target Based Video-music Retrieval With Music Rhythm And Video Optical Flow Information Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, our proposed approach leverages a unified target set to perform video/music pretraining and produces clip-level embeddings to preserve temporal information. |
Tianjun Mao; Shansong Liu; Yunxuan Zhang; Dian Li; Ying Shan; | arxiv-cs.MM | 2023-09-17 |
21 | Estimating Mutual Information for Spike Trains: A Bird Song Example Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Zebra finch are a model animal used in the study of audition. |
Jake Witter; Conor Houghton; | arxiv-cs.IT | 2023-09-14 |
22 | Comparative Assessment of Markov Models and Recurrent Neural Networks for Jazz Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Our study aims to compare the performance of a simple Markov chain model and a recurrent neural network (RNN) model, two popular models for sequence generating tasks, in jazz music improvisation. |
Conrad Hsu; Ross Greer; | arxiv-cs.SD | 2023-09-14 |
23 | SingFake: Singing Voice Deepfake Detection Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose the singing voice deepfake detection task. |
Yongyi Zang; You Zhang; Mojtaba Heydari; Zhiyao Duan; | arxiv-cs.SD | 2023-09-14 |
24 | Undecidability Results and Their Relevance in Modern Music Making Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The study adopts a multidimensional approach, focusing on five key areas: (1) the Turing completeness of Ableton, a widely used digital audio workstation, (2) the undecidability of satisfiability in sound creation utilizing an array of effects, (3) the undecidability of constraints on polymeters in musical compositions, (4) the undecidability of satisfiability in just intonation harmony constraints, and (5) the undecidability of new ordering systems. |
Halley Young; | arxiv-cs.SD | 2023-09-11 |
25 | Exploring Music Genre Classification: Algorithm Analysis and Deployment Architecture Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a study on music genre classification using a combination of Digital Signal Processing (DSP) and Deep Learning (DL) techniques. |
Ayan Biswas; Supriya Dhabal; Palaniandavar Venkateswaran; | arxiv-cs.SD | 2023-09-09 |
26 | A Long-Tail Friendly Representation Framework for Artist and Music Similarity Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes a Long-Tail Friendly Representation Framework (LTFRF) that utilizes neural networks to model the similarity relationship. |
Haoran Xiang; Junyu Dai; Xuchen Song; Furao Shen; | arxiv-cs.SD | 2023-09-08 |
27 | Self-Similarity-Based and Novelty-based Loss for Music Structure Analysis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we propose a supervised approach for the task of music boundary detection. |
Geoffroy Peeters; | arxiv-cs.SD | 2023-09-05 |
28 | FSD: An Initial Chinese Dataset for Fake Song Detection Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Thus, we employ the FSD dataset for the training of ADD models. We subsequently evaluate these models under two scenarios: one with the original songs and another with separated vocal tracks. |
YUANKUN XIE et. al. | arxiv-cs.SD | 2023-09-05 |
29 | MDSC: Towards Evaluating The Style Consistency Between Music and Dance Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose MDSC(Music-Dance-Style Consistency), the first evaluation metric which assesses to what degree the dance moves and music match. |
Zixiang Zhou; Baoyuan Wang; | arxiv-cs.SD | 2023-09-03 |
30 | Towards Contrastive Learning in Music Video Domain Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Contrastive learning is a powerful way of learning multimodal representations across various domains such as image-caption retrieval and audio-visual representation learning. In this work, we investigate if these findings generalize to the domain of music videos. |
Karel Veldkamp; Mariya Hendriksen; Zoltán Szlávik; Alexander Keijser; | arxiv-cs.IR | 2023-09-01 |
31 | Enhancing The Vocal Range of Single-speaker Singing Voice Synthesis with Melody-unsupervised Pre-training Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Based on our previous work, this work proposes a melody-unsupervised multi-speaker pre-training method conducted on a multi-singer dataset to enhance the vocal range of the single-speaker, while not degrading the timbre similarity. |
Shaohuan Zhou; Xu Li; Zhiyong Wu; Ying Shan; Helen Meng; | arxiv-cs.SD | 2023-09-01 |
32 | Sequential Pitch Distributions for Raga Detection Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we attempt to detect the raga using a novel feature to extract sequential or temporal information from an audio sample. |
Vishwaas Narasinh; Senthil Raja G; | arxiv-cs.SD | 2023-08-30 |
33 | A Review of Differentiable Digital Signal Processing for Music & Speech Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The term differentiable digital signal processing describes a family of techniques in which loss function gradients are backpropagated through digital signal processors, facilitating their integration into neural networks. This article surveys the literature on differentiable audio signal processing, focusing on its use in music & speech synthesis. |
Ben Hayes; Jordie Shier; György Fazekas; Andrew McPherson; Charalampos Saitis; | arxiv-cs.SD | 2023-08-29 |
34 | Symbolic & Acoustic: Multi-domain Music Emotion Modeling for Instrumental Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, most prior works were uni-domain and showed weak consistency between arousal modeling performance and valence modeling performance. Based on this background, we designed a multi-domain emotion modeling method for instrumental music that combines symbolic analysis and acoustic analysis. |
Kexin Zhu; Xulong Zhang; Jianzong Wang; Ning Cheng; Jing Xiao; | arxiv-cs.SD | 2023-08-28 |
35 | InstructME: An Instruction Guided Music Edit And Remix Framework with Latent Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we develop InstructME, an Instruction guided Music Editing and remixing framework based on latent diffusion models. |
BING HAN et. al. | arxiv-cs.SD | 2023-08-28 |
36 | Automated Conversion of Music Videos Into Lyric Videos Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, making such videos can be challenging and time-consuming as the lyrics need to be added in synchrony and visual harmony with the video. Informed by prior work and close examination of existing lyric videos, we propose a set of design guidelines to help creators make such videos. |
JIAJU MA et. al. | arxiv-cs.HC | 2023-08-28 |
37 | Fairness Through Domain Awareness: Mitigating Popularity Bias For Music Discovery Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work we explore the intrinsic relationship between music discovery and popularity bias. |
Rebecca Salganik; Fernando Diaz; Golnoosh Farnadi; | arxiv-cs.CY | 2023-08-28 |
38 | Utilizing Mood-Inducing Background Music in Human-Robot Interaction Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: \setcounter{footnote}{2}\footnote{An earlier version of part of the material in this paper appeared originally in the first author’s Ph.D. |
Elad Liebman; Peter Stone; | arxiv-cs.AI | 2023-08-27 |
39 | A Computational Evaluation Framework for Singable Lyric Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present a computational framework for the quantitative evaluation of singable lyric translation, which seamlessly integrates musical, linguistic, and cultural dimensions of lyrics. |
Haven Kim; Kento Watanabe; Masataka Goto; Juhan Nam; | arxiv-cs.CL | 2023-08-25 |
40 | A Comprehensive Survey for Evaluation Methodologies of AI-Generated Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This study aims to comprehensively evaluate the subjective, objective, and combined methodologies for assessing AI-generated music, highlighting the advantages and disadvantages of each approach. |
Zeyu Xiong; Weitao Wang; Jing Yu; Yue Lin; Ziyan Wang; | arxiv-cs.SD | 2023-08-25 |
41 | Emotion-Aligned Contrastive Learning Between Images and Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we address the task of retrieving emotionally-relevant music from image queries by learning an affective alignment between images and music audio. |
Shanti Stewart; Kleanthis Avramidis; Tiantian Feng; Shrikanth Narayanan; | arxiv-cs.MM | 2023-08-24 |
42 | Exploiting Time-Frequency Conformers for Music Audio Enhancement Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Consequently, the necessity for music audio enhancement (referred to as music enhancement from this point onward), involving the transformation of degraded audio recordings into pristine high-quality music, has surged to augment the auditory experience. To address this issue, we propose a music enhancement system based on the Conformer architecture that has demonstrated outstanding performance in speech enhancement tasks. |
Yunkee Chae; Junghyun Koo; Sungho Lee; Kyogu Lee; | arxiv-cs.SD | 2023-08-24 |
43 | Humming2Music: Being A Composer As Long As You Can Humming Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present an automatic music generation system to lower the threshold of creating music. |
Yao Qiu; Jinchao Zhang; Huiying Ren; Yong Shan; Jie Zhou; | ijcai | 2023-08-23 |
44 | Q&A: Query-Based Representation Learning for Multi-Track Symbolic Music Re-Arrangement Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we tackle rearrangement problems via self-supervised learning, in which the mapping styles can be regarded as conditions and controlled in a flexible way. |
Jingwei Zhao; Gus Xia; Ye Wang; | ijcai | 2023-08-23 |
45 | Graph-based Polyphonic Multitrack Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Nonetheless, there is a lack of works that consider graph representations in the context of deep learning systems for music generation. This paper bridges this gap by introducing a novel graph representation for music and a deep Variational Autoencoder that generates the structure and the content of musical graphs separately, one after the other, with a hierarchical architecture that matches the structural priors of music. |
Emanuele Cosenza; Andrea Valenti; Davide Bacciu; | ijcai | 2023-08-23 |
46 | Linear-Sized Spectral Sparsifiers and The Kadison-Singer Problem Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: 182, 327-350 (2015)], has been informally thought of as a strengthening of Batson, Spielman, and Srivastava’s theorem that every undirected graph has a linear-sized spectral sparsifier [SICOMP 41, 1704-1721 (2012)]. We formalize this intuition by using a corollary of the MSS result to derive the existence of spectral sparsifiers with a number of edges linear in its number of vertices for all undirected, weighted graphs. |
Phevos Paschalidis; Ashley Zhuang; | arxiv-cs.DS | 2023-08-23 |
47 | Evaluating Human-AI Interaction Via Usability, User Experience and Acceptance Measures for MMM-C: A Creative AI System for Music Composition Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper reports on a thorough evaluation of the user adoption of the Multi-Track Music Machine (MMM) as a minimal co-creative AI tool for music composers. |
RENAUD BOUGUENG TCHEMEUBE et. al. | ijcai | 2023-08-23 |
48 | LingGe: An Automatic Ancient Chinese Poem-to-Song Generation System Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a novel system, named LingGe ("伶歌" in Chinese), to generate songs for ancient Chinese poems automatically. |
Yong Shan; Jinchao Zhang; Huiying Ren; Yao Qiu; Jie Zhou; | ijcai | 2023-08-23 |
49 | LongDanceDiff: Long-term Dance Generation with Conditional Diffusion Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To enhance the diversity of generated dance motions and mitigate the freezing problem, we introduce a mutual information minimization objective that regularizes the dependency between past and future motions. |
Siqi Yang; Zejun Yang; Zhisheng Wang; | arxiv-cs.CV | 2023-08-23 |
50 | JEPOO: Highly Accurate Joint Estimation of Pitch, Onset and Offset for Music Information Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a highly accurate method for joint estimation of pitch, onset and offset, named JEPOO. |
Haojie Wei; Jun Yuan; Rui Zhang; Yueguo Chen; Gang Wang; | ijcai | 2023-08-23 |
51 | Discrete Diffusion Probabilistic Models for Symbolic Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This work presents the direct generation of Polyphonic Symbolic Music using D3PMs. |
Matthias Plasser; Silvan Peter; Gerhard Widmer; | ijcai | 2023-08-23 |
52 | Music Understanding LLaMA: Advancing Text-to-Music Generation with Question Answering and Captioning Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Text-to-music generation (T2M-Gen) faces a major obstacle due to the scarcity of large-scale publicly available music datasets with natural language captions. To address this, we propose the Music Understanding LLaMA (MU-LLaMA), capable of answering music-related questions and generating captions for music files. |
Shansong Liu; Atin Sakkeer Hussain; Chenshuo Sun; Ying Shan; | arxiv-cs.SD | 2023-08-22 |
53 | MusicJam: Visualizing Music Insights Via Generated Narrative Illustrations Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, in these techniques, the figures are usually pre-selected or statically generated, so they cannot precisely convey insights of different pieces of music. To address this issue, in this paper, we introduce MusicJam, a music visualization system that is able to generate narrative illustrations to represent the insight of the input music. |
CHUER CHEN et. al. | arxiv-cs.HC | 2023-08-22 |
54 | TrOMR:Transformer-Based Polyphonic Optical Music Recognition Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a transformer-based approach with excellent global perceptual capability for end-to-end polyphonic OMR, called TrOMR. |
Yixuan Li; Huaping Liu; Qiang Jin; Miaomiao Cai; Peng Li; | arxiv-cs.CL | 2023-08-18 |
55 | MUSE: Music Recommender System with Shuffle Play Recommendation Enhancement Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Based on our observation that the shuffle play sessions hinder the overall training process of music recommender systems mainly due to the high unique transition rates of shuffle play sessions, we propose a Music Recommender System with Shuffle Play Recommendation Enhancement (MUSE). |
Yunhak Oh; Sukwon Yun; Dongmin Hyun; Sein Kim; Chanyoung Park; | arxiv-cs.IR | 2023-08-18 |
56 | BigWavGAN: A Wave-To-Wave Generative Adversarial Network for Music Super-Resolution Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To unleash the potential of large DNN models in music SR, we propose BigWavGAN, which incorporates Demucs, a large-scale wave-to-wave model, with State-Of-The-Art (SOTA) discriminators and adversarial training strategies. |
Yenan Zhang; Hiroshi Watanabe; | arxiv-cs.SD | 2023-08-12 |
57 | Visual Overviews for Sheet Music Structure Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose different methods for alternative representation and visual augmentation of sheet music that help users gain an overview of general structure, repeating patterns, and the similarity of segments. |
Frank Heyen; Quynh Quang Ngo; Michael Sedlmair; | arxiv-cs.HC | 2023-08-11 |
58 | An Autoethnographic Exploration of XAI in Algorithmic Composition Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper introduces an autoethnographic study of the use of the MeasureVAE generative music XAI model with interpretable latent dimensions trained on Irish folk music. |
Ashley Noel-Hirst; Nick Bryan-Kinns; | arxiv-cs.SD | 2023-08-11 |
59 | DiVa: An Iterative Framework to Harvest More Diverse and Valid Labels from User Comments for Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, current solutions fail to resolve it as they cannot produce diverse enough mappings to make up for the information missed by the gold labels. Based on the observation that such missing information may already be presented in user comments, we propose to study the automated music labeling in an essential but under-explored setting, where the model is required to harvest more diverse and valid labels from the users’ comments given limited gold labels. |
HONGRU LIANG et. al. | arxiv-cs.IR | 2023-08-09 |
60 | JEN-1: Text-Guided Universal Music Generation with Omnidirectional Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper introduces JEN-1, a universal high-fidelity model for text-to-music generation. |
PEIKE LI et. al. | arxiv-cs.SD | 2023-08-09 |
61 | Sudowoodo: A Chinese Lyric Imitation System with Source Lyrics Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce \textbf{\textit{Sudowoodo}}, a Chinese lyrics imitation system that can generate new lyrics based on the text of source lyrics. |
YONGZHU CHANG et. al. | arxiv-cs.CL | 2023-08-08 |
62 | Search Engine and Recommendation System for The Music Industry Built with JinaAI Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Often people face difficulty in searching for a song solely based on the title, hence a solution is proposed to complete a search analysis through a single query input and is matched with the lyrics of the songs present in the database. |
Ishita Gopalakrishnan; Sanjjushri Varshini R; Ponshriharini V; | arxiv-cs.LG | 2023-08-07 |
63 | Amplifying The Music Listening Experience Through Song Comments on Music Streaming Platforms Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, such emotional aspects are often ignored by current platforms, which affects the listeners’ ability to find music that triggers specific personal feelings. To address this gap, this study proposes a novel approach that leverages deep learning methods to capture contextual keywords, sentiments, and induced mechanisms from song comments. |
LONGFEI CHEN et. al. | arxiv-cs.HC | 2023-08-07 |
64 | Bootstrapping Contrastive Learning Enhanced Music Cold-Start Matching Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, there are hardly any studies done on this task. Therefore, in this paper, we will formalize the problem of Music Cold-Start Matching detailedly and give a scheme. |
Xinping Zhao; Ying Zhang; Qiang Xiao; Yuming Ren; Yingchun Yang; | arxiv-cs.IR | 2023-08-05 |
65 | DiffDance: Cascaded Human Motion Diffusion Model for Dance Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Conventional autoregressive methods introduce compounding errors during sampling and struggle to capture the long-term structure of dance sequences. To address these limitations, we present a novel cascaded motion diffusion model, DiffDance, designed for high-resolution, long-form dance generation. |
QIAOSONG QI et. al. | arxiv-cs.GR | 2023-08-05 |
66 | An Interpretable, Flexible, and Interactive Probabilistic Framework for Melody Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Unfortunately, most recent models are practically impossible to interpret or musically fine-tune, as they use deep neural networks with thousands of parameters. We introduce an interpretable, flexible, and interactive model, SchenkComposer, for melody generation that empowers users to be creative in all aspects of the music generation pipeline and allows them to learn from the process. |
Stephen Hahn; Rico Zhu; Simon Mak; Cynthia Rudin; Yue Jiang; | kdd | 2023-08-04 |
67 | Towards Improving Harmonic Sensitivity and Prediction Stability for Singing Melody Extraction Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose an input feature modification and a training objective modification based on two assumptions. |
Keren Shao; Ke Chen; Taylor Berg-Kirkpatrick; Shlomo Dubnov; | arxiv-cs.SD | 2023-08-04 |
68 | The Virtual Drum Circle: Polyrhythmic Music Interactions in Extended Reality Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we designed an extended reality platform to assess a remote, bidirectional polyrhythmic interaction between two players, mediated in real time by their three-dimensional embodied avatars and a shared, virtual drum circle. |
Bavo Van Kerrebroeck; Kristel Crombé; Stéphanie Wilain; Marc Leman; Pieter-Jan Maes; | arxiv-cs.HC | 2023-08-03 |
69 | MusicLDM: Enhancing Novelty in Text-to-Music Generation Using Beat-Synchronous Mixup Strategies Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, to tackle these challenges, we first construct a state-of-the-art text-to-music model, MusicLDM, that adapts Stable Diffusion and AudioLDM architectures to the music domain. We achieve this by retraining the contrastive language-audio pretraining model (CLAP) and the Hifi-GAN vocoder, as components of MusicLDM, on a collection of music data samples. |
KE CHEN et. al. | arxiv-cs.SD | 2023-08-03 |
70 | Music De-limiter Networks Via Sample-wise Gain Inversion Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce music de-limiter networks that estimate uncompressed music from heavily compressed signals. |
Chang-Bin Jeon; Kyogu Lee; | arxiv-cs.SD | 2023-08-02 |
71 | LP-MusicCaps: LLM-Based Pseudo Music Captioning Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Despite its importance, researchers face challenges due to the costly and time-consuming collection process of existing music-language datasets, which are limited in size. To address this data scarcity issue, we propose the use of large language models (LLMs) to artificially generate the description sentences from large-scale tag datasets. |
SeungHeon Doh; Keunwoo Choi; Jongpil Lee; Juhan Nam; | arxiv-cs.SD | 2023-07-30 |
72 | Towards A New Interface for Music Listening: A User Experience Study on YouTube Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We also propose wireframes of a video streaming service for better audio-visual music listening in two stages: search and listening. By these wireframes, we offer practical solutions to enhance user satisfaction with YouTube for music listening. |
Ahyeon Choi; Eunsik Shin; Haesun Joung; Joongseek Lee; Kyogu Lee; | arxiv-cs.HC | 2023-07-27 |
73 | DisCover: Disentangled Music Representation Learning for Cover Song Identification Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we set the goal of disentangling version-specific and version-invariant factors, which could make it easier for the model to learn invariant music representations for unseen query songs. |
JIAHAO XUN et. al. | sigir | 2023-07-25 |
74 | When The Music Stops: Tip-of-the-Tongue Retrieval for Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present a study of Tip-of-the-tongue (ToT) retrieval for music, where a searcher is trying to find an existing music entity, but is unable to succeed as they cannot accurately recall important identifying information. |
Samarth Bhargav; Anne Schuth; Claudia Hauff; | sigir | 2023-07-25 |
75 | Music Genre Classification with ResNet and Bi-GRU Using Visual Spectrograms Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: On the other hand, deep learning classification architectures like the traditional Convolutional Neural Networks (CNN) are effective in capturing the spatial hierarchies but struggle to capture the temporal dynamics inherent in music data. To address these challenges, this study proposes a novel approach using visual spectrograms as input, and propose a hybrid model that combines the strength of the Residual neural Network (ResNet) and the Gated Recurrent Unit (GRU). |
Junfei Zhang; | arxiv-cs.SD | 2023-07-20 |
76 | From West to East: Who Can Understand The Music of The Others Better? Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: At the same time, the vast majority of these models have been trained on Western pop/rock music and related styles. This leads to research questions on whether these models can be used to learn representations for different music cultures and styles, or whether we can build similar music audio embedding models trained on data from different cultures or styles. |
Charilaos Papaioannou; Emmanouil Benetos; Alexandros Potamianos; | arxiv-cs.SD | 2023-07-19 |
77 | JAZZVAR: A Dataset of Variations Found Within Solo Piano Performances of Jazz Standards for Music Overpainting Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we outline the curation process for obtaining and sorting the repertoire, the pipeline for creating the Original and Variation pairs, and our analysis of the dataset. |
Eleanor Row; Jingjing Tang; George Fazekas; | arxiv-cs.SD | 2023-07-18 |
78 | ProgGP: From GuitarPro Tablature Neural Generation To Progressive Metal Production Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We extend this work by fine-tuning a pre-trained Transformer model on ProgGP, a custom dataset of 173 progressive metal songs, for the purposes of creating compositions from that genre through a human-AI partnership. |
Jackson Loth; Pedro Sarmento; CJ Carr; Zack Zukowski; Mathieu Barthet; | arxiv-cs.SD | 2023-07-11 |
79 | Optimizing Feature Extraction for Symbolic Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper presents a comprehensive investigation of existing feature extraction tools for symbolic music and contrasts their performance to determine the set of features that best characterizes the musical style of a given music score. |
Federico Simonetta; Ana Llorens; Martín Serrano; Eduardo García-Portugués; Álvaro Torrente; | arxiv-cs.SD | 2023-07-11 |
80 | On The Effectiveness of Speech Self-supervised Learning for Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Nevertheless, research exploring the effectiveness of applying speech SSL models to music recordings has been limited. We explore the music adaption of SSL with two distinctive speech-related models, data2vec1.0 and Hubert, and refer to them as music2vec and musicHuBERT, respectively. |
YINGHAO MA et. al. | arxiv-cs.SD | 2023-07-11 |
81 | VampNet: Music Generation Via Masked Acoustic Token Modeling Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce VampNet, a masked acoustic token modeling approach to music synthesis, compression, inpainting, and variation. |
Hugo Flores Garcia; Prem Seetharaman; Rithesh Kumar; Bryan Pardo; | arxiv-cs.SD | 2023-07-10 |
82 | Unsupervised Melody-to-Lyrics Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a method for generating high-quality lyrics without training on any aligned melody-lyric data. |
YUFEI TIAN et. al. | acl | 2023-07-08 |
83 | Emotion-Guided Music Accompaniment Generation Based on Variational Autoencoder Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Existing models struggle to effectively characterize human emotions within neural network models while composing music. To address this issue, we propose the use of an easy-to-represent emotion flow model, the Valence/Arousal Curve, which allows for the compatibility of emotional information within the model through data transformation and enhances interpretability of emotional factors by utilizing a Variational Autoencoder as the model structure. |
Qi Wang; Shubing Zhang; Li Zhou; | arxiv-cs.SD | 2023-07-08 |
84 | Songs Across Borders: Singable and Controllable Neural Lyric Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper bridges the singability quality gap by formalizing lyric translation into a constrained translation problem, converting theoretical guidance and practical techniques from translatology literature to prompt-driven NMT approaches, exploring better adaptation methods, and instantiating them to an English-Chinese lyric translation system. |
Longshen Ou; Xichu Ma; Min-Yen Kan; Ye Wang; | acl | 2023-07-08 |
85 | UniLG: A Unified Structure-aware Framework for Lyrics Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a unified structure-aware lyrics generation framework named UniLG. |
TAO QIAN et. al. | acl | 2023-07-08 |
86 | LaunchpadGPT: Language Model As Music Visualization Designer on Launchpad Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Launchpad is a musical instrument that allows users to create and perform music by pressing illuminated buttons. To assist and inspire the design of the Launchpad light effect, and provide a more accessible approach for beginners to create music visualization with this instrument, we proposed the LaunchpadGPT model to generate music visualization designs on Launchpad automatically. |
Siting Xu; Yunlong Tang; Feng Zheng; | arxiv-cs.SD | 2023-07-07 |
87 | Track Mix Generation on Music Streaming Services Using Transformers Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper introduces Track Mix, a personalized playlist generation system released in 2022 on the music streaming service Deezer. |
WALID BENDADA et. al. | arxiv-cs.IR | 2023-07-06 |
88 | LOAF-M2L: Joint Learning of Wording and Formatting for Singable Melody-to-Lyric Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper bridges the singability gap with a novel approach to generating singable lyrics by jointly Learning wOrding And Formatting during Melody-to-Lyric training (LOAF-M2L). |
Longshen Ou; Xichu Ma; Ye Wang; | arxiv-cs.CL | 2023-07-05 |
89 | Musif: A Python Package for Symbolic Music Feature Extraction Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we introduce musif, a Python package that facilitates the automatic extraction of features from symbolic music scores. |
Ana Llorens; Federico Simonetta; Martín Serrano; Álvaro Torrente; | arxiv-cs.SD | 2023-07-03 |
90 | Audio Embeddings As Teachers for Music Classification Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we integrate the ideas of transfer learning and feature-based knowledge distillation and systematically investigate using pre-trained audio embeddings as teachers to guide the training of low-complexity student networks. |
Yiwei Ding; Alexander Lerch; | arxiv-cs.SD | 2023-06-30 |
91 | Predicting Music Hierarchies with A Graph-Based Neural Decoder Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper describes a data-driven framework to parse musical sequences into dependency trees, which are hierarchical structures used in music cognition research and music analysis. |
Francesco Foscarin; Daniel Harasim; Gerhard Widmer; | arxiv-cs.SD | 2023-06-29 |
92 | RMVPE: A Robust Model for Vocal Pitch Estimation in Polyphonic Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: As a result, the performance of vocal pitch estimation is affected by the music source separation models. To address this issue and directly extract vocal pitches from polyphonic music, we propose a robust model named RMVPE. |
Haojie Wei; Xueke Cao; Tangpeng Dan; Yueguo Chen; | arxiv-cs.SD | 2023-06-27 |
93 | Fine-Grained Position Helps Memorizing More, A Novel Music Compound Transformer Model with Feature Interaction Fusion Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Therefore, in this work, we propose an improved compound Transformer model for music understanding. Specifically, we propose an attribute embedding fusion module and a novel position encoding scheme with absolute-relative consideration. |
Zuchao Li; Ruhan Gong; Yineng Chen; Kehua Su; | aaai | 2023-06-26 |
94 | Melody: A Platform for Linked Open Data Visualisation and Curated Storytelling Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this article we present MELODY, a web interface for authoring data stories based on Linked Open Data. |
Giulia Renda; Marilena Daquino; Valentina Presutti; | arxiv-cs.HC | 2023-06-26 |
95 | Mining and Applying Composition Knowledge of Dance Moves for Style-Concentrated Dance Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Without the stylized prior knowledge, these approaches are not promising to generate controllable style or diverse moves for each dance style, nor new dances complying with stylized knowledge. To address this issue, we propose a novel music-to-dance generation framework guided by style embedding, considering both input music and stylized dancing knowledge. |
Xinjian Zhang; Su Yang; Yi Xu; Weishan Zhang; Longwen Gao; | aaai | 2023-06-26 |
96 | A Domain-Knowledge-Inspired Music Embedding Space and A Novel Attention Mechanism for Symbolic Music Modeling Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose the Fundamental Music Embedding (FME) for symbolic music based on a bias-adjusted sinusoidal encoding within which both the absolute and the relative attributes can be embedded and the fundamental musical properties (e.g., translational invariance) are explicitly preserved. |
Zixun Guo; Jaeyong Kang; Dorien Herremans; | aaai | 2023-06-26 |
97 | DISCO-10M: A Large-Scale Music Dataset Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: With DISCO-10M, we aim to democratize and facilitate new research to help advance the development of novel machine learning models for music. |
Luca A. Lanzendörfer; Florian Grötschla; Emil Funke; Roger Wattenhofer; | arxiv-cs.SD | 2023-06-23 |
98 | From Ontology Design to User-centred Interfaces for Music Heritage Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this article we investigate the bridge between ontology design and UI/UX design methodologies to assist designers in prototyping web applications for information seeking purposes. |
Giulia Renda; Marco Grasso; Marilena Daquino; | arxiv-cs.HC | 2023-06-22 |
99 | Toward Leveraging Pre-Trained Self-Supervised Frontends for Automatic Singing Voice Understanding Tasks: Three Case Studies Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: By fine-tuning these models for the target tasks, comparable performance to conventional supervised learning can be achieved with limited training data. Therefore, in this paper, we investigate the effectiveness of SSL models for various singing voice recognition tasks. |
Yuya Yamamoto; | arxiv-cs.SD | 2023-06-22 |
100 | Knowledge-based Multimodal Music Similarity Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The aim of this research is to develop a fully explainable and interpretable system that can provide end-users with more control and understanding of music similarity and classification systems. |
Andrea Poltronieri; | arxiv-cs.SD | 2023-06-21 |
101 | Multitrack Music Transcription with A Time-Frequency Perceiver Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel deep neural network architecture, Perceiver TF, to model the time-frequency representation of audio input for multitrack transcription. |
Wei-Tsung Lu; Ju-Chiang Wang; Yun-Ning Hung; | arxiv-cs.SD | 2023-06-19 |
102 | MARBLE: Music Audio Representation Benchmark for Universal Evaluation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In the era of extensive intersection between art and Artificial Intelligence (AI), such as image generation and fiction co-creation, AI for music remains relatively nascent, particularly in music understanding. This is evident in the limited work on deep music representations, the scarcity of large-scale datasets, and the absence of a universal and community-driven benchmark. |
RUIBIN YUAN et. al. | arxiv-cs.SD | 2023-06-18 |
103 | CoverHunter: Cover Song Identification with Refined Attention and Alignments Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a novel system named CoverHunter that overcomes the shortcomings of existing detection schemes by exploring richer features with refined attention and alignments. |
Feng Liu; Deyi Tuo; Yinan Xu; Xintong Han; | arxiv-cs.SD | 2023-06-15 |
104 | Language-Guided Music Recommendation for Video Via Prompt Analogies Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a method to recommend music for an input video while allowing a user to guide music selection with free-form natural language. |
Daniel McKee; Justin Salamon; Josef Sivic; Bryan Russell; | arxiv-cs.CV | 2023-06-15 |
105 | Exploring Isolated Musical Notes As Pre-training Data for Predominant Instrument Recognition in Polyphonic Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work presents our efforts toward building a robust end-to-end instrument recognition system for polyphonic multi-instrument music. |
Lifan Zhong; Erica Cooper; Junichi Yamagishi; Nobuaki Minematsu; | arxiv-cs.SD | 2023-06-15 |
106 | Sound Demixing Challenge 2023 Music Demixing Track Technical Report: TFC-TDF-UNet V3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this report, we present our award-winning solutions for the Music Demixing Track of Sound Demixing Challenge 2023. |
Minseok Kim; Jun Hyung Lee; Soonyoung Jung; | arxiv-cs.SD | 2023-06-15 |
107 | Anticipatory Music Transformer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce anticipation: a method for constructing a controllable generative model of a temporal point process (the event process) conditioned asynchronously on realizations of a second, correlated process (the control process). |
John Thickstun; David Hall; Chris Donahue; Percy Liang; | arxiv-cs.SD | 2023-06-14 |
108 | Empirical Measurement of Aesthetic Experience of Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present an affordable, wearable, easy-to-carry device to measure phenomenological goosebumps intensity on our skin with respect to real-time data using IoT devices (Raspberry pi 3, model B). |
Abhishek Gupta; C. M. Markan; | arxiv-cs.HC | 2023-06-13 |
109 | Video-to-Music Recommendation Using Temporal Alignment of Segments Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a novel approach to significantly improve the system’s performance using structure-aware recommendation. |
Laure Prétet; Gaël Richard; Clément Souchier; Geoffroy Peeters; | arxiv-cs.MM | 2023-06-12 |
110 | Everybody Compose: Deep Beats To Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This project presents a deep learning approach to generate monophonic melodies based on input beats, allowing even amateurs to create their own music compositions. |
Conghao Shen; Violet Z. Yao; Yixin Liu; | arxiv-cs.SD | 2023-06-09 |
111 | Simple and Controllable Music Generation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce MusicGen, a single Language Model (LM) that operates over several streams of compressed discrete music representation, i.e., tokens. |
JADE COPET et. al. | arxiv-cs.SD | 2023-06-08 |
112 | SANGEET: A XML Based Open Dataset for Research in Hindustani Sangeet Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: The dataset is intended to provide the ground truth information for music information research tasks, thereby supporting several data-driven analysis from a machine learning perspective. We present the usefulness of the dataset by demonstrating its application on music information retrieval using XQuery, visualization through Omenad rendering system. |
Chandan Misra; Swarup Chattopadhyay; | arxiv-cs.SD | 2023-06-07 |
113 | Emotion-Conditioned Melody Harmonization with Hierarchical Variational Autoencoder Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Meanwhile, the variability of harmonies generated by previous methods is insufficient. To solve these problems, we propose a novel LSTM-based Hierarchical Variational Auto-Encoder (LHVAE) to investigate the influence of emotional conditions on melody harmonization, while improving the quality of generated harmonies and capturing the abundant variability of chord progressions. |
Shulei Ji; Xinyu Yang; | arxiv-cs.SD | 2023-06-06 |
114 | Music Mode: Transforming Robot Movement Into Music Increases Likability and Perceived Intelligence Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present Music Mode, a novel mapping between a robot’s joint motions and sounds, programmed by artists and engineers to make the robot generate music as it moves. |
Catie Cuan; Emre Fisher; Allison Okamura; Tom Engbersen; | arxiv-cs.RO | 2023-06-05 |
115 | Controllable Lyrics-to-Melody Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Controllability of generative models enables human interaction with models to generate desired contents, which is especially important in music generation tasks towards human-centered AI that can facilitate musicians in creative activities. To address these issues, we propose a controllable lyrics-to-melody generation network, ConL2M, which is able to generate realistic melodies from lyrics in user-desired musical style. |
Zhe Zhang; Yi Yu; Atsuhiro Takasu; | arxiv-cs.SD | 2023-06-05 |
116 | LyricSIM: A Novel Dataset and Benchmark for Similarity Detection in Spanish Song LyricS Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present a new dataset and benchmark tailored to the task of semantic similarity in song lyrics. |
ALEJANDRO BENITO-SANTOS et. al. | arxiv-cs.CL | 2023-06-02 |
117 | Q&A: Query-Based Representation Learning for Multi-Track Symbolic Music Re-Arrangement Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we tackle rearrangement problems via self-supervised learning, in which the mapping styles can be regarded as conditions and controlled in a flexible way. |
Jingwei Zhao; Gus Xia; Ye Wang; | arxiv-cs.SD | 2023-06-02 |
118 | Learning Music Sequence Representation from Text Supervision Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To excavate better MUsic SEquence Representation from labeled audio, we propose a novel text-supervision pre-training method, namely MUSER. |
TIANYU CHEN et. al. | arxiv-cs.SD | 2023-05-31 |
119 | Transfer Learning for Underrepresented Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper investigates a combinational creativity approach to transfer learning to improve the performance of deep neural network-based models for music generation on out-of-distribution (OOD) genres. |
Anahita Doosti; Matthew Guzdial; | arxiv-cs.LG | 2023-05-31 |
120 | MuseCoco: Generating Symbolic Music from Text Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose MuseCoco, which generates symbolic music from text descriptions with musical attributes as the bridge to break down the task into text-to-attribute understanding and attribute-to-music generation stages. |
PEILING LU et. al. | arxiv-cs.SD | 2023-05-31 |
121 | MERT: Acoustic Music Understanding Model with Large-Scale Self-supervised Training Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To address this research gap, we propose an acoustic Music undERstanding model with large-scale self-supervised Training (MERT), which incorporates teacher models to provide pseudo labels in the masked language modelling (MLM) style acoustic pre-training. |
YIZHI LI et. al. | arxiv-cs.SD | 2023-05-31 |
122 | Unsupervised Melody-to-Lyric Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a method for generating high-quality lyrics without training on any aligned melody-lyric data. |
YUFEI TIAN et. al. | arxiv-cs.CL | 2023-05-30 |
123 | Efficient Neural Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present MeLoDy (M for music; L for LM; D for diffusion), an LM-guided diffusion model that generates music audios of state-of-the-art quality meanwhile reducing 95.7% or 99.6% forward passes in MusicLM, respectively, for sampling 10s or 30s music. |
MAX W. Y. LAM et. al. | arxiv-cs.SD | 2023-05-25 |
124 | LoopBoxes — Evaluation of A Collaborative Accessible Digital Musical Instrument Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper describes the evaluation of the instrument in the form of a pilot study during a music festival in Berlin, Germany, as well as a case study with children and music teachers in a SEN school setting. |
Andreas Förster; Alarith Uhde; Mathias Komesker; Christina Komesker; Irina Schmidt; | arxiv-cs.HC | 2023-05-24 |
125 | MIDI-Draw: Sketching to Control Melody Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We describe a proof-of-principle implementation of a system for drawing melodies that abstracts away from a note-level input representation via melodic contours. |
Tashi Namgyal; Peter Flach; Raul Santos-Rodriguez; | arxiv-cs.SD | 2023-05-19 |
126 | GETMusic: Generating Any Music Tracks with A Unified Representation and Diffusion Framework Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Previous works fail to address this need due to inherent constraints in music representations and model architectures. To address this need, we propose a unified representation and diffusion framework named GETMusic (`GET’ stands for GEnerate music Tracks), which includes a novel music representation named GETScore, and a diffusion model named GETDiff. |
ANG LV et. al. | arxiv-cs.SD | 2023-05-18 |
127 | RMSSinger: Realistic-Music-Score Based Singing Voice Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Furthermore, these manual annotation destroys the regularity of note durations in music scores, making fine-grained music scores inconvenient for composing. To tackle these challenges, we propose RMSSinger, the first RMS-SVS method, which takes realistic music scores as input, eliminating most of the tedious manual annotation and avoiding the aforementioned inconvenience. |
JINZHENG HE et. al. | arxiv-cs.SD | 2023-05-17 |
128 | Music-Driven Group Choreography Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present AIOZ-GDANCE, a new largescale dataset for music-driven group dance generation. |
NHAT LE et. al. | cvpr | 2023-05-17 |
129 | EDGE: Editable Dance Generation From Music IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we introduce Editable Dance GEneration (EDGE), a state-of-the-art method for editable dance generation that is capable of creating realistic, physically-plausible dances while remaining faithful to the input music. |
Jonathan Tseng; Rodrigo Castellon; Karen Liu; | cvpr | 2023-05-17 |
130 | SongDriver2: Real-time Emotion-based Music Arrangement with Soft Transition Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Existing studies mainly focus on achieving emotion real-time fit, while the issue of soft transition remains understudied, affecting the overall emotional coherence of the music. In this paper, we propose SongDriver2 to address this balance. |
ZIHAO WANG et. al. | arxiv-cs.SD | 2023-05-13 |
131 | Music Rearrangement Using Hierarchical Segmentation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a method for automatically rearranging music recordings that takes into account the hierarchical structure of the recording. |
Christos Plachouras; Marius Miron; | arxiv-cs.SD | 2023-05-12 |
132 | Unsupervised Melody-Guided Lyrics Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose to generate pleasantly listenable lyrics without training on melody-lyric aligned data. |
YUFEI TIAN et. al. | arxiv-cs.AI | 2023-05-12 |
133 | V2Meow: Meowing to The Visual Beat Via Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel approach called V2Meow that can generate high-quality music audio that aligns well with the visual semantics of a diverse range of video input types. |
KUN SU et. al. | arxiv-cs.SD | 2023-05-11 |
134 | Learn to Sing By Listening: Building Controllable Virtual Singer By Unsupervised Learning from Voice Recordings Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a framework that could digitize a person’s voice by simply listening to the clean voice recordings of any content in a fully unsupervised manner and predict singing voices even only using speaking recordings. |
Wei Xue; Yiwen Wang; Qifeng Liu; Yike Guo; | arxiv-cs.SD | 2023-05-09 |
135 | Exploring Softly Masked Language Modelling for Controllable Symbolic Music Generation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: This document presents some early explorations of applying Softly Masked Language Modelling (SMLM) to symbolic music generation. SMLM can be seen as a generalisation of masked … |
Nicolas Jonason; Bob L. T. Sturm; | arxiv-cs.SD | 2023-05-05 |
136 | LooPy: A Research-Friendly Mix Framework for Music Information Retrieval on Electronic Dance Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, music genres like electronic dance music (EDM) has always been relatively less investigated compared to others. Considering its wide range of applications, we present a Python package for automated EDM audio generation as an infrastructure for MIR for EDM songs, to mitigate the difficulty of acquiring labelled data. |
Xinyu Li; | arxiv-cs.SD | 2023-05-01 |
137 | Transfer of Knowledge Among Instruments in Automatic Music Transcription Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work shows how to employ easily generated synthesized audio data produced by software synthesizers to train a universal model. |
Michał Leś; Michał Woźniak; | arxiv-cs.SD | 2023-04-30 |
138 | Improving Music Genre Classification from Multi-modal Properties of Music and Genre Correlations Perspective Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In addition, as genres normally co-occur in a music track, it is desirable to capture and model the genre correlations to improve the performance of multi-label music genre classification. To solve these issues, we present a novel multi-modal method leveraging audio-lyrics contrastive loss and two symmetric cross-modal attention, to align and fuse features from audio and lyrics. |
G. Ru; X. Zhang; J. Wang; N. Cheng; J. Xiao; | icassp | 2023-04-27 |
139 | Audio Quality Assessment of Vinyl Music Collections Using Self-Supervised Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we show that the self-supervised learning (SSL) model wav2vec 2.0 can be successfully used to predict the perceived audio quality of archive music collections. |
A. Ragano; E. Benetos; A. Hines; | icassp | 2023-04-27 |
140 | Controllable Music Inpainting with Mixed-Level and Disentangled Representation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we contribute a controllable inpainting model by combining the high expressivity of mixed-level, disentangled music representations and the strong predictive power of masked language modeling. |
S. Wei; Z. Wang; W. Gao; G. Xia; | icassp | 2023-04-27 |
141 | Music Rearrangement Using Hierarchical Segmentation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a method for automatically rearranging music recordings that takes into account the hierarchical structure of the recording. |
C. Plachouras; M. Miron; | icassp | 2023-04-27 |
142 | Contrastive Self-Supervised Learning for Automated Multi-Modal Dance Performance Assessment Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: A fundamental challenge of analyzing human motion is to effectively represent human movements both spatially and temporally. We propose a contrastive self-supervised strategy to tackle this challenge. |
Y. Zhong; F. Zhang; Y. Demiris; | icassp | 2023-04-27 |
143 | Enhancing The Vocal Range of Single-Speaker Singing Voice Synthesis with Melody-Unsupervised Pre-Training Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Based on our previous work, this work proposes a melody-unsupervised multi-speaker pretraining method conducted on a multi-singer dataset to enhance the vocal range of the single-speaker, while not degrading the timbre similarity. |
S. Zhou; X. Li; Z. Wu; Y. Shan; H. Meng; | icassp | 2023-04-27 |
144 | SSVMR: Saliency-Based Self-Training for Video-Music Retrieval IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel saliency-based self-training framework, which is termed SSVMR. |
X. Cheng; Z. Zhu; H. Li; Y. Li; Y. Zou; | icassp | 2023-04-27 |
145 | An Attention-Based Approach to Hierarchical Multi-Label Music Instrument Classification Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: For the effective joint training in the multi-label setting, we propose two methods to model the connection between fine- and coarse-level tags, where one uses rule-based grouped max-pooling, the other one uses the attention mechanism obtained in a data-driven manner. |
Z. ZHONG et. al. | icassp | 2023-04-27 |
146 | Phoneix: Acoustic Feature Processing Strategy for Enhanced Singing Pronunciation With Phoneme Distribution Predictor Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose an acoustic feature processing strategy, named PHONEix, with a phoneme distribution predictor, to alleviate the gap between the music score and the singing voice, which can be easily adopted in different SVS systems. |
Y. Wu; J. Shi; T. Qian; D. Gao; Q. Jin; | icassp | 2023-04-27 |
147 | Music Mixing Style Transfer: A Contrastive Learning Approach to Disentangle Audio Effects Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose an end-to-end music mixing style transfer system that converts the mixing style of an input multitrack to that of a reference song. |
J. KOO et. al. | icassp | 2023-04-27 |
148 | HIPI: A Hierarchical Performer Identification Model Based on Symbolic Representation of Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we apply a Recurrent Neural Network (RNN) model to classify the most likely music performers from their interpretative styles. |
S. R. Mahmud Rafee; G. Fazekas; G. Wiggins; | icassp | 2023-04-27 |
149 | Textless Speech-to-Music Retrieval Using Emotion Similarity Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce a framework that recommends music based on the emotions of speech. |
S. Doh; M. Won; K. Choi; J. Nam; | icassp | 2023-04-27 |
150 | Continuous Interaction with A Smart Speaker Via Low-Dimensional Embeddings of Dynamic Hand Pose Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a new continuous interaction strategy with visual feedback of hand pose and mid-air gesture recognition and control for a smart music speaker, which utilizes only 2 video frames to recognize gestures. |
S. Xu; C. Kaul; X. Ge; R. Murray-Smith; | icassp | 2023-04-27 |
151 | Pre-Training Strategies Using Contrastive Learning and Playlist Information for Music Classification and Similarity Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we investigate an approach that relies on contrastive learning and music metadata as a weak source of supervision to train music representation models. |
P. ALONSO-JIMÉNEZ et. al. | icassp | 2023-04-27 |
152 | Unsupervised Vocal Dereverberation with Diffusion-Based Generative Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, recent supervised dereverberation methods may fail because they rely on sufficiently diverse and numerous pairs of reverberant observations and retrieved data for training in order to be generalizable to unseen observations during inference. To resolve these problems, we propose an unsupervised method that can remove a general kind of artificial reverb for music without requiring pairs of data for training. |
K. SAITO et. al. | icassp | 2023-04-27 |
153 | Self-Supervised Hierarchical Metrical Structure Modeling Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a novel method to model hierarchical metrical structures for both symbolic music and audio signals in a self-supervised manner with minimal domain knowledge. |
J. Jiang; G. Xia; | icassp | 2023-04-27 |
154 | Multitrack Music Transcription with A Time-Frequency Perceiver Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel deep neural network architecture, Perceiver TF, to model the time-frequency representation of audio input for multitrack transcription. |
W. . -T. Lu; J. -C. Wang; Y. . -N. Hung; | icassp | 2023-04-27 |
155 | Phonation Mode Detection in Singing: A Singer Adapted Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we define the phonation mode detection (PMD) problem, which entails the prediction of phonation mode labels as well as their onset and offset timestamps. |
Y. Wang; W. Wei; Y. Wang; | icassp | 2023-04-27 |
156 | Direct Position Determination with One-Bit Signal for Multiple Targets Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The traditional direct position determination (DPD) for multiple targets usually requires transmitting raw data to the fusion center (FC), which occupies large transmission bandwidth and hardware resource. To solve this problem, we adopt one-bit analog-to-digital converters (ADCs) for a distributed subarray (DS) system, and propose an one-bit DPD method with multiple signal classification (1-bit DPD-MUSIC). |
L. NI et. al. | icassp | 2023-04-27 |
157 | Self-Supervised Representations for Singing Voice Conversion Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we circumvent disentanglement training and propose a new model that leverages ASR fine-tuned self-supervised representations as inputs to a HiFi-GAN neural vocoder for singing voice conversion. |
T. JAYASHANKAR et. al. | icassp | 2023-04-27 |
158 | Deep Root Music Algorithm for Data-Driven Doa Estimation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work we propose Deep Root-MUSIC (DR-MUSIC); a data-driven DoA estimator which augments Root-MUSIC with a deep neural network applied to the empirical autocorrelation of the input. |
D. H. Shmuel; J. P. Merkofer; G. Revach; R. J. G. van Sloun; N. Shlezinger; | icassp | 2023-04-27 |
159 | Knowledge-Graph Augmented Music Representation for Genre Classification Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose KGenre, a knowledge-embedded music representation learning framework for improved genre classification. |
H. DING et. al. | icassp | 2023-04-27 |
160 | Toward Universal Text-To-Music Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper introduces effective design choices for text-to-music retrieval systems. |
S. Doh; M. Won; K. Choi; J. Nam; | icassp | 2023-04-27 |
161 | MAID: A Conditional Diffusion Model for Long Music Audio Inpainting Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, the information about these segments may differ significantly from the original. To solve this problem, we propose MAID (Music Audio Inpainting DDPM), a model for music audio inpainting based on DDPM (Denoising Diffusion Probability Model). |
K. Liu; W. Gan; C. Yuan; | icassp | 2023-04-27 |
162 | Bytecover3: Accurate Cover Song Identification On Short Queries Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we upgrade the previous ByteCover systems to ByteCover3 that utilizes local features to further improve the identification performance of short music queries. |
X. DU et. al. | icassp | 2023-04-27 |
163 | Diffroll: Diffusion-Based Generative Music Transcription with Unsupervised Pretraining Capability Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we propose a novel generative approach, DiffRoll, to tackle automatic music transcription (AMT). |
K. W. CHEUK et. al. | icassp | 2023-04-27 |
164 | Compose & Embellish: Well-Structured Piano Performance Generation Via A Two-Stage Approach Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Observing the above, we devise a two-stage Transformer-based framework that Composes a lead sheet first, and then Embellishes it with accompaniment and expressive touches. |
S. -L. Wu; Y. -H. Yang; | icassp | 2023-04-27 |
165 | Jazznet: A Dataset of Fundamental Piano Patterns for Music Audio Machine Learning Research Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper introduces the jazznet Dataset, a dataset of fundamental jazz piano music patterns for developing machine learning (ML) algorithms in music information retrieval (MIR). |
T. Adegbija; | icassp | 2023-04-27 |
166 | Audio-Driven Facial Landmark Generation in Violin Performance Using 3DCNN Network with Self Attention Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we compile a violin soundtrack and facial expression dataset (VSFE) for modeling facial expressions in violin performance. |
T. -W. Lin; C. -L. Liu; L. Su; | icassp | 2023-04-27 |
167 | Chord-Conditioned Melody Harmonization With Controllable Harmonicity Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: Melody harmonization has long been closely associated with chorales composed by Johann Sebastian Bach. Previous works rarely emphasised chorale generation conditioned on chord … |
S. Wu; X. Li; M. Sun; | icassp | 2023-04-27 |
168 | TrOMR:Transformer-Based Polyphonic Optical Music Recognition Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a transformer-based approach with excellent global perceptual capability for end-to-end polyphonic OMR, called TrOMR. |
Y. Li; H. Liu; Q. Jin; M. Cai; P. Li; | icassp | 2023-04-27 |
169 | Multitrack Music Transformer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a new multitrack music representation that allows a diverse set of instruments while keeping a short sequence length. |
H. -W. Dong; K. Chen; S. Dubnov; J. McAuley; T. Berg-Kirkpatrick; | icassp | 2023-04-27 |
170 | Error Analysis of Convolutional Beamspace Algorithms Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, theoretical MSE of CBS is given when MUSIC or root-MUSIC is used. |
P. -C. Chen; P. P. Vaidyanathan; | icassp | 2023-04-27 |
171 | GTN-Bailando: Genre Consistent Long-Term 3D Dance Generation Based on Pre-Trained Genre Token Network Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In addition, the correlation between the dance genre and the music has not been investigated. To address these issues, we propose a genre-consistent dance generation framework, GTN-Bailando. |
H. ZHUANG et. al. | icassp | 2023-04-27 |
172 | On The Role of Visual Context in Enriching Music Representations Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this study, we propose VCMR – Video-Conditioned Music Representations, a contrastive learning framework that learns music representations from audio and the accompanying music videos. |
K. Avramidis; S. Stewart; S. Narayanan; | icassp | 2023-04-27 |
173 | Pop2Piano : Pop Audio-Based Piano Cover Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present Pop2Piano, a Transformer network that generates piano covers given waveforms of pop music. |
J. Choi; K. Lee; | icassp | 2023-04-27 |
174 | GTN-Bailando: Genre Consistent Long-Term 3D Dance Generation Based on Pre-trained Genre Token Network Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In addition, the correlation between the dance genre and the music has not been investigated. To address these issues, we propose a genre-consistent dance generation framework, GTN-Bailando. |
HAOLIN ZHUANG et. al. | arxiv-cs.SD | 2023-04-25 |
175 | Pre-Training Strategies Using Contrastive Learning and Playlist Information for Music Classification and Similarity Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we investigate an approach that relies on contrastive learning and music metadata as a weak source of supervision to train music representation models. |
PABLO ALONSO-JIMÉNEZ et. al. | arxiv-cs.SD | 2023-04-24 |
176 | An Order-Complexity Model for Aesthetic Quality Assessment of Homophony Music Performance Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In order to guide the generation task of AI music performance, and to improve the performance effect of human performers, this paper uses Birkhoff’s aesthetic measure to propose a method of objective measurement of beauty. |
XIN JIN et. al. | arxiv-cs.SD | 2023-04-22 |
177 | CLaMP: Contrastive Language-Music Pre-training for Cross-Modal Symbolic Music Information Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce CLaMP: Contrastive Language-Music Pre-training, which learns cross-modal representations between natural language and symbolic music using a music encoder and a text encoder trained jointly with a contrastive loss. |
Shangda Wu; Dingyao Yu; Xu Tan; Maosong Sun; | arxiv-cs.SD | 2023-04-21 |
178 | From Words to Music: A Study of Subword Tokenization Techniques in Symbolic Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: As Transformer models become increasingly popular in symbolic music-related studies, it is imperative to investigate the efficacy of subword tokenization in the symbolic music domain. In this paper, we explore subword tokenization techniques, such as byte-pair encoding (BPE), in symbolic music generation and its impact on the overall structure of generated songs. |
Adarsh Kumar; Pedro Sarmento; | arxiv-cs.SD | 2023-04-18 |
179 | Generative Disco: Text-to-Video Generation for Music Visualization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce Generative Disco, a generative AI system that helps generate music visualizations with large language models and text-to-video generation. |
Vivian Liu; Tao Long; Nathan Raw; Lydia Chilton; | arxiv-cs.HC | 2023-04-17 |
180 | Adapting Meter Tracking Models to Latin American Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Furthermore, when studying these music traditions, models are not required to generalize to multiple mainstream music genres but to perform well in more constrained, homogeneous conditions. In this work, we investigate simple yet effective strategies to adapt beat and downbeat tracking models to two different Latin American music traditions and analyze the feasibility of these adaptations in real-world applications concerning the data and computational requirements. |
Lucas S. Maia; Martín Rocamora; Luiz W. P. Biscainho; Magdalena Fuentes; | arxiv-cs.SD | 2023-04-14 |
181 | Self-supervised Auxiliary Loss for Metric Learning in Music Similarity-based Retrieval and Auto-tagging Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we propose a model that builds on the self-supervised learning approach to address the similarity-based retrieval challenge by introducing our method of metric learning with a self-supervised auxiliary loss. |
Taketo Akama; Hiroaki Kitano; Katsuhiro Takematsu; Yasushi Miyajima; Natalia Polouliakh; | arxiv-cs.SD | 2023-04-14 |
182 | AffectMachine-Classical: A Novel System for Generating Affective Classical Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work introduces a new music generation system, called AffectMachine-Classical, that is capable of generating affective Classic music in real-time. |
Kat R. Agres; Adyasha Dash; Phoebe Chua; | arxiv-cs.SD | 2023-04-10 |
183 | Adoption of AI Technology in The Music Mixing Workflow: An Investigation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Through semi-structured interviews, a questionnaire-based study, and analyzing web forums, the study confirms three user groups comprising amateurs, pro-ams, and professionals. |
Soumya Sai Vanka; Maryam Safi; Jean-Baptiste Rolland; George Fazekas; | arxiv-cs.HC | 2023-04-06 |
184 | TM2D: Bimodality Driven 3D Dance Generation Via Music-Text Integration Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a novel task for generating 3D dance movements that simultaneously incorporate both text and music modalities. |
KEHONG GONG et. al. | arxiv-cs.CV | 2023-04-05 |
185 | Pac-HuBERT: Self-Supervised Music Source Separation Via Primitive Auditory Clustering and Hidden-Unit BERT Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a self-supervised learning framework for music source separation inspired by the HuBERT speech representation model. |
Ke Chen; Gordon Wichern; François G. Germain; Jonathan Le Roux; | arxiv-cs.SD | 2023-04-04 |
186 | High-Throughput Vector Similarity Search in Knowledge Graphs Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we explore vector similarity search in the context of Knowledge Graphs (KGs). |
JASON MOHONEY et. al. | arxiv-cs.DB | 2023-04-04 |
187 | Adaptive Background Music for A Fighting Game: A Multi-Instrument Volume Modulation Approach Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents our work to enhance the background music (BGM) in DareFightingICE by adding an adaptive BGM. |
Ibrahim Khan; Thai Van Nguyen; Chollakorn Nimpattanavong; Ruck Thawonmas; | arxiv-cs.SD | 2023-03-28 |
188 | Translate The Beauty in Songs: Jointly Learning to Align Melody and Translate Lyrics Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose Lyrics-Melody Translation with Adaptive Grouping (LTAG), a holistic solution to automatic song translation by jointly modeling lyrics translation and lyrics-melody alignment. |
CHENGXI LI et. al. | arxiv-cs.CL | 2023-03-27 |
189 | MusicFace: Music-driven Expressive Singing Face Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: It is still an interesting and challenging problem to synthesize a vivid and realistic singing face driven by music signal. In this paper, we present a method for this task with natural motions of the lip, facial expression, head pose, and eye states. |
PENGFEI LIU et. al. | arxiv-cs.GR | 2023-03-24 |
190 | A-MuSIC: An Adaptive Ensemble System For Visual Place Recognition In Changing Environments Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: As a result, VPR systems combining multiple techniques achieve more reliable VPR performance in changing environments, at the cost of higher computational loads. Addressing this shortcoming, we propose an adaptive VPR system dubbed Adaptive Multi-Self Identification and Correction (A-MuSIC). |
Bruno Arcanjo; Bruno Ferrarini; Michael Milford; Klaus D. McDonald-Maier; Shoaib Ehsan; | arxiv-cs.CV | 2023-03-24 |
191 | Symbolic Music Structure Analysis with Graph Representations and Changepoint Detection Methods Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In the past, there have been several works that attempt to segment music into the audio and symbolic domains, however, the identification and segmentation of the music structure at different levels is still an open research problem in this area. In this work we propose three methods, two of which are novel graph-based algorithms that aim to segment symbolic music by its form or structure: Norm, G-PELT and G-Window. |
Carlos Hernandez-Olivan; Sonia Rubio Llamas; Jose R. Beltran; | arxiv-cs.SD | 2023-03-24 |
192 | VMCML: Video and Music Matching Via Cross-Modality Lifting Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a content-based system for matching video and background music. |
Yi-Shan Lee; Wei-Cheng Tseng; Fu-En Wang; Min Sun; | arxiv-cs.CV | 2023-03-22 |
193 | ByteCover3: Accurate Cover Song Identification on Short Queries Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we upgrade the previous ByteCover systems to ByteCover3 that utilizes local features to further improve the identification performance of short music queries. |
XINGJIAN DU et. al. | arxiv-cs.SD | 2023-03-21 |
194 | Controllable Ancient Chinese Lyrics Generation Based on Phrase Prototype Retrieving Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a novel method for generating ancient Chinese lyrics (Song Ci), a type of ancient lyrics that involves precise control of song structure. |
Li Yi; | arxiv-cs.CL | 2023-03-20 |
195 | Textless Speech-to-Music Retrieval Using Emotion Similarity Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce a framework that recommends music based on the emotions of speech. |
SeungHeon Doh; Minz Won; Keunwoo Choi; Juhan Nam; | arxiv-cs.SD | 2023-03-18 |
196 | Tollywood Emotions: Annotation of Valence-Arousal in Telugu Song Lyrics Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a new manually annotated dataset of Telugu songs’ lyrics collected from Spotify with valence and arousal annotated on a discrete scale. |
R Guru Ravi Shanker; B Manikanta Gupta; BV Koushik; Vinoo Alluri; | arxiv-cs.CL | 2023-03-16 |
197 | Generating Symbolic Music Using Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, a diffusion model that uses a binomial prior distribution to generate piano rolls is proposed. |
Lilac Atassi; | arxiv-cs.SD | 2023-03-15 |
198 | PHONEix: Acoustic Feature Processing Strategy for Enhanced Singing Pronunciation with Phoneme Distribution Predictor Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose an acoustic feature processing strategy, named PHONEix, with a phoneme distribution predictor, to alleviate the gap between the music score and the singing voice, which can be easily adopted in different SVS systems. |
Yuning Wu; Jiatong Shi; Tao Qian; Dongji Gao; Qin Jin; | arxiv-cs.SD | 2023-03-15 |
199 | Improving Music Genre Classification from Multi-Modal Properties of Music and Genre Correlations Perspective Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In addition, as genres normally co-occur in a music track, it is desirable to capture and model the genre correlations to improve the performance of multi-label music genre classification. To solve these issues, we present a novel multi-modal method leveraging audio-lyrics contrastive loss and two symmetric cross-modal attention, to align and fuse features from audio and lyrics. |
Ganghui Ru; Xulong Zhang; Jianzong Wang; Ning Cheng; Jing Xiao; | arxiv-cs.SD | 2023-03-14 |
200 | DiffuseRoll: Multi-track Multi-category Music Generation Based on Diffusion Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel approach for multi-track, multi-attribute symphonic music generation using the diffusion model. |
Hongfei Wang; | arxiv-cs.SD | 2023-03-14 |
201 | A Human Subject Study of Named Entity Recognition (NER) in Conversational Music Recommendation Queries Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Our goal was to learn about the task to guide the design of better evaluation methods and NER algorithms. |
Elena V. Epure; Romain Hennequin; | arxiv-cs.CL | 2023-03-13 |
202 | Hybrid Y-Net Architecture for Singing Voice Separation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Overall, our work presents a promising approach for improving the accuracy and efficiency of music source separation. |
Rashen Fernando; Pamudu Ranasinghe; Udula Ranasinghe; Janaka Wijayakulasooriya; Pantaleon Perera; | arxiv-cs.SD | 2023-03-05 |
203 | Decoding and Visualising Intended Emotion in An Expressive Piano Performance Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Expert musicians can mould a musical piece to convey specific emotions that they intend to communicate. In this paper, we place a mid-level features based music emotion model in this performer-to-listener communication scenario, and demonstrate via a small visualisation music emotion decoding in real time. |
Shreyan Chowdhury; Gerhard Widmer; | arxiv-cs.SD | 2023-03-03 |
204 | Continuous Interaction with A Smart Speaker Via Low-dimensional Embeddings of Dynamic Hand Pose Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a new continuous interaction strategy with visual feedback of hand pose and mid-air gesture recognition and control for a smart music speaker, which utilizes only 2 video frames to recognize gestures. |
Songpei Xu; Chaitanya Kaul; Xuri Ge; Roderick Murray-Smith; | arxiv-cs.HC | 2023-02-28 |
205 | A Comparative Analysis Of Latent Regressor Losses For Singing Voice Conversion Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose an alternative loss component in a loss function that is otherwise well-established among VC tasks, which has been shown to improve our model’s SVC performance. |
Brendan O’Connor; Simon Dixon; | arxiv-cs.SD | 2023-02-27 |
206 | Multi-Modality in Music: Predicting Emotion in Music from High-Level Audio Features and Lyrics Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper aims to test whether a multi-modal approach for music emotion recognition (MER) performs better than a uni-modal one on high-level song features and lyrics. |
Tibor Krols; Yana Nikolova; Ninell Oldenburg; | arxiv-cs.SD | 2023-02-26 |
207 | Cybersickness, Cognition, & Motor Skills: The Effects of Music, Gender, and Gaming Experience Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, 39 participants performed an assessment four times, once before the rides (baseline), and then once after each ride (3 rides). |
Panagiotis Kourtesis; Josie Linnell; Rayaan Amir; Ferran Argelaguet; Sarah E. MacPherson; | arxiv-cs.HC | 2023-02-25 |
208 | Computational Creativity: Compose The Music for A Movie Using Only Its Automatically Extracted Brightness Curve Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we are interested in how the computer can be used as a tool to promote creativity in a musical composition. |
Felipe Ariani; Marcelo Caetano; Javier Elipe Gimeno; Ivan Magrin-Chagnolleau; | arxiv-cs.MM | 2023-02-20 |
209 | SSVMR: Saliency-based Self-training for Video-Music Retrieval IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel saliency-based self-training framework, which is termed SSVMR. |
Xuxin Cheng; Zhihong Zhu; Hongxiang Li; Yaowei Li; Yuexian Zou; | arxiv-cs.MM | 2023-02-18 |
210 | An Attention-based Approach to Hierarchical Multi-label Music Instrument Classification Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: For the effective joint training in the multi-label setting, we propose two methods to model the connection between fine- and coarse-level tags, where one uses rule-based grouped max-pooling, the other one uses the attention mechanism obtained in a data-driven manner. |
ZHI ZHONG et. al. | arxiv-cs.SD | 2023-02-16 |
211 | Mixing Levels — A Rock Music Spirit Level App Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this manuscript, we present Mixing Levels, a spirit level sonification based on music mixing. |
Tim Ziemer; | arxiv-cs.MM | 2023-02-16 |
212 | Jazznet: A Dataset of Fundamental Piano Patterns for Music Audio Machine Learning Research Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper introduces the jazznet Dataset, a dataset of fundamental jazz piano music patterns for developing machine learning (ML) algorithms in music information retrieval (MIR). |
Tosiron Adegbija; | arxiv-cs.SD | 2023-02-16 |
213 | Unsupervised Classification to Improve The Quality of A Bird Song Recording Dataset Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Although labelling functions were introduced to expedite strong label assignment, their classification performance remains mostly unknown. To address this issue and reduce label noise (wrong label assignment) in large bird song datasets, we introduce a data-centric novel labelling function composed of three successive steps: 1) time-frequency sound unit segmentation, 2) feature computation for each sound unit, and 3) classification of each sound unit as bird song or noise with either an unsupervised DBSCAN algorithm or the supervised BirdNET neural network. |
Félix Michaud; Jérôme Sueur; Maxime Le Cesne; Sylvain Haupert; | arxiv-cs.LG | 2023-02-15 |
214 | GTR-CTRL: Instrument and Genre Conditioning for Guitar-Focused Music Generation with Transformers Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we use the DadaGP dataset for guitar tab music generation, a corpus of over 26k songs in GuitarPro and token formats. |
PEDRO SARMENTO et. al. | arxiv-cs.SD | 2023-02-10 |
215 | ERNIE-Music: Text-to-Waveform Music Generation with Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Nevertheless, the direct synthesis of music waveforms from unrestricted textual prompts remains a relatively underexplored domain. In response to this lacuna, this paper introduces a pioneering contribution in the form of a text-to-waveform music generation model, underpinned by the utilization of diffusion models. |
PENGFEI ZHU et. al. | arxiv-cs.SD | 2023-02-09 |
216 | Incorporating Total Variation Regularization in The Design of An Intelligent Query By Humming System Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we incorporate Total Variation Regularization (TVR) to denoise queries. |
Shivangi Ranjan; Vishal Srivastava; | arxiv-cs.SD | 2023-02-09 |
217 | Goniometers Are A Powerful Acoustic Feature for Music Information Retrieval Tasks Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: Goniometers, also known as Phase Scopes or Vector Scopes, are audio metering tools that help music producers and mixing engineers monitor spatial aspects of a music mix, such as … |
Tim Ziemer; | arxiv-cs.SD | 2023-02-02 |
218 | FIGARO: Controllable Music Generation Using Learned and Expert Features Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we release FIGARO, a Transformer-based conditional model trained to generate symbolic music based on a sequence of high-level control codes. |
Dimitri von Rütte; Luca Biggio; Yannic Kilcher; Thomas Hofmann; | iclr | 2023-02-01 |
219 | A Nearly-Optimal Bound for Fast Regression with $\ell_\infty$ Guarantee Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: One popular approach for solving such $\ell_2$ regression problem is via sketching: picking a structured random matrix $S\in \mathbb{R}^{m\times n}$ with $m\ll n$ and $SA$ can be quickly computed, solve the “sketched” regression problem $\arg\min_{x\in \mathbb{R}^d} \|SAx-Sb\|_2$. In this paper, we show that in order to obtain such $\ell_\infty$ guarantee for $\ell_2$ regression, one has to use sketching matrices that are dense. |
Zhao Song; Mingquan Ye; Junze Yin; Lichen Zhang; | arxiv-cs.DS | 2023-02-01 |
220 | Channel Reconstruction-Aided MUSIC Algorithms for Joint AoA&AoD Estimation in MIMO Systems Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: This letter discusses the issue of joint angle-of-arrival (AoA) and angle-of-departure (AoD) estimation for multi-input multi-output (MIMO) arrays, so as to conceive a novel … |
Teng Ma; Yue Xiao; Xia Lei; | IEEE Wireless Communications Letters | 2023-02-01 |
221 | An Analysis of Classification Approaches for Hit Song Prediction Using Engineered Metadata Features with Lyrics and Audio Features Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This study aims to improve the prediction result of the top 10 hits among Billboard Hot 100 songs using more alternative metadata, including song audio features provided by Spotify, song lyrics, and novel metadata-based features (title topic, popularity continuity and genre class). |
Mengyisong Zhao; Morgan Harvey; David Cameron; Frank Hopfgartner; Valerie J. Gillet; | arxiv-cs.IR | 2023-01-31 |
222 | An Comparative Analysis of Different Pitch and Metrical Grid Encoding Methods in The Task of Sequential Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a integrated analysis of the influence of two low-level feature, pitch and meter, on the performance of a token-based sequential music generation model. |
Yuqiang Li; Shengchen Li; George Fazekas; | arxiv-cs.SD | 2023-01-30 |
223 | Large Music Recommendation Studies for Small Teams Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Running live music recommendation studies without direct industry partnerships can be a prohibitively daunting task, especially for small teams. In order to help future researchers interested in such evaluations, we present a number of struggles we faced in the process of generating our own such evaluation system alongside potential solutions. |
Kyle Robinson; Dan Brown; | arxiv-cs.HC | 2023-01-30 |
224 | SingSong: Generating Musical Accompaniments from Singing IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present SingSong, a system that generates instrumental music to accompany input vocals, potentially offering musicians and non-musicians alike an intuitive new way to create music featuring their own voice. |
CHRIS DONAHUE et. al. | arxiv-cs.SD | 2023-01-29 |
225 | Automated Arrangements of Multi-Part Music for Sets of Monophonic Instruments Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we study the problem of automating music arrangements for music pieces written for monophonic instruments or voices. |
Matthew Mccloskey; Gabrielle Curcio; Amulya Badineni; Kevin Mcgrath; Dimitris Papamichail; | arxiv-cs.SD | 2023-01-27 |
226 | Byte Pair Encoding for Symbolic Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Until now, the proposed tokenizations are based on small vocabularies describing the note attributes and time events, resulting in fairly long token sequences. In this paper, we show how Byte Pair Encoding (BPE) can improve the results of deep learning models while improving its performances. |
Nathan Fradet; Jean-Pierre Briot; Fabien Chhel; Amal El Fallah Seghrouchni; Nicolas Gutowski; | arxiv-cs.LG | 2023-01-27 |
227 | Moûsai: Text-to-Music Generation with Long-Context Latent Diffusion IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We develop a cascading latent diffusion approach that can generate multiple minutes of high-quality stereo music at 48kHz from textual descriptions. |
Flavio Schneider; Zhijing Jin; Bernhard Schölkopf; | arxiv-cs.CL | 2023-01-27 |
228 | Reconstruction of Blade Tip-timing Signals Based on The MUSIC Algorithm IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View |
ZHIBO LIU et. al. | Mechanical Systems and Signal Processing | |
229 | MusicLM: Generating Music From Text IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce MusicLM, a model generating high-fidelity music from text descriptions such as a calming violin melody backed by a distorted guitar riff. |
ANDREA AGOSTINELLI et. al. | arxiv-cs.SD | 2023-01-26 |
230 | Wish You Were Here: Mental and Physiological Effects of Remote Music Collaboration in Mixed Reality Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: With face-to-face music collaboration being severely limited during the recent pandemic, mixed reality technologies and their potential to provide musicians a feeling of being there with their musical partner can offer tremendous opportunities. In order to assess this potential, we conducted a laboratory study in which musicians made music together in real-time while simultaneously seeing their jamming partner’s mixed reality point cloud via a head-mounted display and compared mental effects such as flow, affect, and co-presence to an audio-only baseline. |
RUBEN SCHLAGOWSKI et. al. | arxiv-cs.HC | 2023-01-23 |
231 | Dance2MIDI: Dance-driven Multi-instruments Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Based on our proposed dataset, we introduce a multi-instruments MIDI generation framework (Dance2MIDI) conditioned on dance video. |
Bo Han; Yi Ren; Yuheng Li; | arxiv-cs.MM | 2023-01-22 |
232 | Beyond The Big Five Personality Traits for Music Recommendation Systems Related Papers Related Patents Related Grants Related Venues Related Experts View |
Mariusz Kleć; Alicja Wieczorkowska; K. Szklanny; Włodzimierz Strus; | EURASIP Journal on Audio, Speech, and Music Processing | 2023-01-19 |
233 | SpotHitPy: A Study For ML-Based Song Hit Prediction Using Spotify Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we approached the Hit Song Prediction problem, which aims to predict which songs will become Billboard hits. |
Ioannis Dimolitsas; Spyridon Kantarelis; Afroditi Fouka; | arxiv-cs.SD | 2023-01-19 |
234 | A Hybrid Deep Learning Approach for Classification of Music Genres Using Wavelet and Spectrogram Analysis Related Papers Related Patents Related Grants Related Venues Related Experts View |
K. K. Jena; Sourav Kumar Bhoi; S. Mohapatra; Sambit Bakshi; | Neural Computing and Applications | 2023-01-19 |
235 | Msanii: High Fidelity Music Synthesis on A Shoestring Budget Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present Msanii, a novel diffusion-based model for synthesizing long-context, high-fidelity music efficiently. |
Kinyugo Maina; | arxiv-cs.SD | 2023-01-16 |
236 | An Order-Complexity Model for Aesthetic Quality Assessment of Symbolic Homophony Music Scores Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Based on Birkhoff’s aesthetic measure, this paper proposes an objective quantitative evaluation method for homophony music score aesthetic quality assessment. |
XIN JIN et. al. | arxiv-cs.SD | 2023-01-14 |
237 | In BLOOM: Creativity and Affinity in Artificial Lyrics and Art Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We apply a large multilingual language model (BLOOM-176B) in open-ended generation of Chinese song lyrics, and evaluate the resulting lyrics for coherence and creativity using human reviewers. |
Evan Crothers; Herna Viktor; Nathalie Japkowicz; | arxiv-cs.CL | 2023-01-13 |
238 | WuYun: Exploring Hierarchical Skeleton-guided Melody Generation Using Knowledge-enhanced Deep Learning Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Here, we present WuYun, a knowledge-enhanced deep learning architecture for improving the structure of generated melodies, which first generates the most structurally important notes to construct a melodic skeleton and subsequently infills it with dynamically decorative notes into a full-fledged melody. |
KEJUN ZHANG et. al. | arxiv-cs.SD | 2023-01-11 |
239 | Why People Skip Music? On Predicting Music Skips Using Deep Reinforcement Learning Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we study the utility of users’ historical data for the task of sequentially predicting users’ skipping behaviour. |
Francesco Meggetto; Crawford Revie; John Levine; Yashar Moshfeghi; | arxiv-cs.IR | 2023-01-10 |
240 | Multi-Genre Music Transformer — Composing Full Length Musical Piece Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The objective of the project is to implement a Multi-Genre Transformer which learns to produce music pieces through more adaptive learning process involving more challenging task where genres or form of the composition is also considered. |
Abhinav Kaushal Keshari; | arxiv-cs.SD | 2023-01-06 |
241 | Multimodal Lyrics-Rhythm Matching Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This is likely due to challenges such as audio misalignment, inaccuracies in syllabic identification, and most importantly, the need for cross-disciplinary knowledge. To address this lack of research, we propose a novel multimodal lyrics-rhythm matching approach in this paper that specifically matches key components of lyrics and music with each other without any language limitations. |
Callie C. Liao; Duoduo Liao; Jesse Guessford; | arxiv-cs.SD | 2023-01-06 |
242 | Validity in Music Information Research Experiments Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this article, we review the subject of validity in general, considering the four major types of validity from a key reference: Shadish et al. 2002. |
Bob L. T. Sturm; Arthur Flexer; | arxiv-cs.SD | 2023-01-04 |
243 | Language Models Are Drummers: Drum Composition with Natural Language Pre-Training Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Evaluating generated music is a challenging task, more so is evaluating drum grooves with little precedence in literature. Hence, we propose a tailored structural evaluation method and analyze drum grooves produced by GPT3 compared to those played by human professionals, exposing the strengths and weaknesses of such generation by language-to-music transfer. |
Li Zhang; Chris Callison-Burch; | arxiv-cs.SD | 2023-01-03 |
244 | PMG-Net: Persian Music Genre Classification Using Deep Neural Networks Related Papers Related Patents Related Grants Related Venues Related Experts View |
N. Farajzadeh; Nima Sadeghzadeh; Mahdi Hashemzadeh; | Entertain. Comput. | 2023-01-01 |
245 | Affective Idiosyncratic Responses to Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Despite consensus that idiosyncratic factors play a key role in regulating how listeners emotionally respond to music, precisely measuring the marginal effects of these variables has proved challenging. To address this gap, we develop computational methods to measure affective responses to music from over 403M listener comments on a Chinese social music platform. |
Sky CH-Wang; Evan Li; Oliver Li; Smaranda Muresan; Zhou Yu; | emnlp | 2022-12-30 |
246 | TeleMelody: Lyric-to-Melody Generation with A Template-Based Two-Stage Method IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we develop TeleMelody, a two-stage lyric-to-melody generation system with music template (e.g., tonality, chord progression, rhythm pattern, and cadence) to bridge the gap between lyrics and melodies (i.e., the system consists of a lyric-to-template module and a template-to-melody module). |
ZEQIAN JU et. al. | emnlp | 2022-12-30 |
247 | Data-Efficient Playlist Captioning With Musical and Linguistic Knowledge Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose PlayNTell, a data-efficient multi-modal encoder-decoder model for automatic playlist captioning. |
Giovanni Gabbolini; Romain Hennequin; Elena Epure; | emnlp | 2022-12-30 |
248 | Universality of Preference Behaviors in Online Music-listener Bipartite Networks: A Big Data Analysis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We combine the methods from complex networks theory and information sciences within the context of Big Data analysis to unveil statistical patterns and community structures underlying the formation and evolution of musical preference behaviors. |
Xiao-Pu Han; Fen Lin; Jonathan J. H. Zhu; Tarik Hadzibeganovic; | arxiv-cs.SI | 2022-12-26 |
249 | Navigation Method Enhancing Music Listening Experience By Stimulating Both Neck Sides with Modulated Music Vibration Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a method that stimulates musical vibration (generated from and synchronized with musical signals), modulated by the direction and distance to the target, on both sides of a user’s neck with Hapbeat, a necklace-type haptic device.We conducted three experiments to confirm that the proposed method can achieve both haptic navigation and enhance the music-listening experience.Experiment 1 consisted of conducting a questionnaire survey to examine the effect of stimulating musical vibrations.Experiment 2 evaluated the accuracy (deg) of users’ ability to adjust their direction toward a target using the proposed method.Experiment 3 examined the ability of four different navigation methods by performing navigation tasks in a virtual environment.The results of the experiments showed that stimulating musical vibration enhanced the music-listening experience, and that the proposed method is able to provide sufficient information to guide the users: accuracy in identifying directions was about 20 deg, participants reached the target in all navigation tasks, and in about 80% of all trials participants reached the target using the shortest route.Furthermore, the proposed method succeeded in conveying distance information, and Hapbeat can be combined with conventional navigation methods without interfering with music listening. |
Yusuke Yamazaki; Shoichi Hasegawa; | arxiv-cs.HC | 2022-12-26 |
250 | Generating Music with Sentiment Using Transformer-GANs Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a generative model of symbolic music conditioned by data retrieved from human sentiment. |
Pedro Neves; Jose Fornari; João Florindo; | arxiv-cs.SD | 2022-12-21 |
251 | ALCAP: Alignment-Augmented Music Captioner Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose to explicitly learn the multimodal alignment through contrastive learning. |
ZIHAO HE et. al. | arxiv-cs.SD | 2022-12-21 |
252 | Polytopic Analysis of Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present this model, which we call polytopic analysis of music, along with a new opensource dedicated toolbox called MusicOnPolytopes (in Python). |
Axel Marmoret; Jérémy E. Cohen; Frédéric Bimbot; | arxiv-cs.SD | 2022-12-21 |
253 | Redefining Relationships in Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: While these tools can have the potential to empower creativity, they may fundamentally redefine relationships between stakeholders, to the benefit of some and the detriment of others. In this position paper, we argue that these tools will fundamentally reshape our music culture, with profound effects (for better and for worse) on creators, consumers and the commercial enterprises that often connect them. |
CHRISTIAN DETWEILER et. al. | arxiv-cs.CY | 2022-12-13 |
254 | A Comparison of Audio Preprocessing Techniques and Deep Learning Algorithms for Raga Recognition Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we used two different digital audio signal processing techniques to preprocess audio samples of Carnatic classical ragas that were then processed by various Deep Learning models. |
Devayani Hebbar; Vandana Jagtap; | arxiv-cs.SD | 2022-12-10 |
255 | TRBLLmaker — Transformer Reads Between Lyrics Lines Maker Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present a generative model that offers implicit meanings for several lines of a song. |
Mor Ventura; Michael Toker; | arxiv-cs.CL | 2022-12-09 |
256 | MAP-Music2Vec: A Simple and Effective Baseline for Self-Supervised Music Audio Representation Learning Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The deep learning community has witnessed an exponen-tially growing interest in self-supervised learning (SSL). However, it still remains unexplored how to build a framework for … |
YIZHI LI et. al. | ArXiv | 2022-12-05 |
257 | Melody Transcription Via Generative Pre-training Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: A key challenge in melody transcription is building methods which can handle broad audio containing any number of instrument ensembles and musical styles – existing strategies work well for some melody instruments or styles but not all. To confront this challenge, we leverage representations from Jukebox (Dhariwal et al. 2020), a generative model of broad music audio, thereby improving performance on melody transcription by $20$% relative to conventional spectrogram features. |
Chris Donahue; John Thickstun; Percy Liang; | arxiv-cs.SD | 2022-12-04 |
258 | Sonus Texere! Automated Dense Soundtrack Construction for Books Using Movie Adaptations Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose the first fully automatic method to build a dense soundtrack for books, which can play high-quality instrumental music for the entirety of the reading duration. |
Jaidev Shriram; Makarand Tapaswi; Vinoo Alluri; | arxiv-cs.SD | 2022-12-02 |
259 | A Domain-Knowledge-Inspired Music Embedding Space and A Novel Attention Mechanism for Symbolic Music Modeling Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose the Fundamental Music Embedding (FME) for symbolic music based on a bias-adjusted sinusoidal encoding within which both the absolute and the relative attributes can be embedded and the fundamental musical properties (e.g., translational invariance) are explicitly preserved. |
Z. Guo; J. Kang; D. Herremans; | arxiv-cs.SD | 2022-12-02 |
260 | Assessing The Impact of Music Recommendation Diversity on Listeners: A Longitudinal Study Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present the results of a 12-week longitudinal user study wherein the participants, 110 subjects from Southern Europe, received on a daily basis Electronic Music (EM) diversified recommendations. |
Lorenzo Porcaro; Emilia Gómez; Carlos Castillo; | arxiv-cs.HC | 2022-12-01 |
261 | Proceedings of The 2nd International Workshop on Reading Music Systems Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The International Workshop on Reading Music Systems (WoRMS) is a workshop that tries to connect researchers who develop systems for reading music, such as in the field of Optical … |
Jorge Calvo-Zaragoza; Alexander Pacha; | arxiv-cs.CV | 2022-12-01 |
262 | Proceedings of The 3rd International Workshop on Reading Music Systems Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The International Workshop on Reading Music Systems (WoRMS) is a workshop that tries to connect researchers who develop systems for reading music, such as in the field of Optical … |
Jorge Calvo-Zaragoza; Alexander Pacha; | arxiv-cs.CV | 2022-12-01 |
263 | MuSFA: Improving Music Structural Function Analysis with Partially Labeled Data Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose to repurpose a public dataset, HookTheory Lead Sheet Dataset (HLSD), to improve the performance. |
Ju-Chiang Wang; Jordan B. L. Smith; Yun-Ning Hung; | arxiv-cs.SD | 2022-11-28 |
264 | OK Computer Analysis: An Audio Corpus Study of Radiohead Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In the present work, a corpus of Radiohead songs across their career from 1992 to 2017 are subjected to automated audio analysis. |
Nick Collins; | arxiv-cs.SD | 2022-11-28 |
265 | SongRewriter: A Chinese Song Rewriting System with Controllable Content and Rhyme Scheme Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Although lyrics generation has achieved significant progress in recent years, it has limited practical applications because the generated lyrics cannot be performed without composing compatible melodies. In this work, we bridge this practical gap by proposing a song rewriting system which rewrites the lyrics of an existing song such that the generated lyrics are compatible with the rhythm of the existing melody and thus singable. |
Yusen Sun; Liangyou Li; Qun Liu; Dit-Yan Yeung; | arxiv-cs.CL | 2022-11-27 |
266 | BEAMERS: Brain-Engaged, Active Music-based Emotion Regulation System Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The work shows that: (1) a novel music-based emotion regulation system with a commercial EEG device is designed without employing deterministic emotion recognition models for daily usage; (2) the system considers users’ variant emotions towards the same song, and by which calculate user’s emotion instability and it is in accordance with Big Five Personality Test; (3) the system supports different emotion regulation styles with users’ designation of desired emotion variation, and achieves an accuracy of over $0.85$ with 2-seconds EEG data; (4) people feel easier to report their emotion variation comparing with absolute emotional states, and would accept a more delicate music recommendation system for emotion regulation according to the questionnaire. |
Jiyang Li; Wei Wang; Kratika Bhagtani; Yincheng Jin; Zhanpeng Jin; | arxiv-cs.HC | 2022-11-26 |
267 | EEGLog: Lifelogging EEG Data When You Listen to Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: With the development of wearable EEG devices, we introduce a novel modality of logging EEG data while listening to music, and bring up the idea of the neural-centric way of life with the designed data analysis application named EEGLog. |
JIYANG LI et. al. | arxiv-cs.HC | 2022-11-26 |
268 | Toward Universal Text-to-Music Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper introduces effective design choices for text-to-music retrieval systems. |
SeungHeon Doh; Minz Won; Keunwoo Choi; Juhan Nam; | arxiv-cs.IR | 2022-11-26 |
269 | Proceedings of The 4th International Workshop on Reading Music Systems Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The International Workshop on Reading Music Systems (WoRMS) is a workshop that tries to connect researchers who develop systems for reading music, such as in the field of Optical … |
Jorge Calvo-Zaragoza; Alexander Pacha; Elona Shatri; | arxiv-cs.CV | 2022-11-23 |
270 | A Dataset for Greek Traditional and Folk Music: Lyra Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, apart from presenting the dataset in detail, we propose a baseline deep-learning classification approach to recognize the involved musicological attributes. |
Charilaos Papaioannou; Ioannis Valiantzas; Theodoros Giannakopoulos; Maximos Kaliakatsos-Papakostas; Alexandros Potamianos; | arxiv-cs.SD | 2022-11-21 |
271 | Exploring The Efficacy of Pre-trained Checkpoints in Text-to-Music Generation Task Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we carry out the first study of generating complete and semantically consistent symbolic music scores from text descriptions, and explore the efficacy of using publicly available checkpoints (i.e., BERT, GPT-2, and BART) for natural language processing in the task of text-to-music generation. |
Shangda Wu; Maosong Sun; | arxiv-cs.SD | 2022-11-21 |
272 | EDGE: Editable Dance Generation From Music IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we introduce Editable Dance GEneration (EDGE), a state-of-the-art method for editable dance generation that is capable of creating realistic, physically-plausible dances while remaining faithful to the input music. |
Jonathan Tseng; Rodrigo Castellon; C. Karen Liu; | arxiv-cs.SD | 2022-11-19 |
273 | ComMU: Dataset for Combinatorial Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce combinatorial music generation, a new task to create varying background music based on given conditions. |
LEE HYUN et. al. | arxiv-cs.SD | 2022-11-17 |
274 | Video-Music Retrieval:A Dual-Path Cross-Modal Network Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a method to recommend background music for videos. |
Xin Gu; Yinghua Shen; Chaohui Lv; | arxiv-cs.MM | 2022-11-16 |
275 | A Review of Intelligent Music Generation Systems Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Some of the available literature reviews lack a systematic benchmark of generative models and are traditional and conservative in their perspective, resulting in a vision of the future development of the field that is not deeply integrated with the current rapid scientific progress. In this paper, we conduct a comprehensive survey and analysis of recent intelligent music generation techniques,provide a critical discussion, explicitly identify their respective characteristics, and present them in a general table. |
ZIYI ZHAO et. al. | arxiv-cs.SD | 2022-11-16 |
276 | Show Me The Instruments: Musical Instrument Retrieval from Mixture Audio Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To search the musical instrument samples or virtual instruments that make one’s desired sound, music producers use their ears to listen and compare each instrument sample in their collection, which is time-consuming and inefficient. In this paper, we call this task as Musical Instrument Retrieval and propose a method for retrieving desired musical instruments using reference music mixture as a query. |
KYUNGSU KIM et. al. | arxiv-cs.SD | 2022-11-15 |
277 | Music Instrument Classification Reprogrammed Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: The performance of approaches to Music Instrument Classification, a popular task in Music Information Retrieval, is often impacted and limited by the lack of availability of annotated data for training. We propose to address this issue with reprogramming, a technique that utilizes pre-trained deep and complex neural networks originally targeting a different task by modifying and mapping both the input and output of the pre-trained model. |
Hsin-Hung Chen; Alexander Lerch; | arxiv-cs.SD | 2022-11-15 |
278 | YM2413-MDB: A Multi-Instrumental FM Video Game Music Dataset with Emotion Annotations Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose YM2413-MDB, an 80s FM video game music dataset with multi-label emotion annotations. |
EUNJIN CHOI et. al. | arxiv-cs.SD | 2022-11-14 |
279 | Music Similarity Calculation of Individual Instrumental Sounds Using Metric Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Towards the development of a more flexible music recommendation system, we propose a music similarity calculation method that focuses on individual instrumental sound sources in a musical piece. |
Yuka Hashizume; Li Li; Tomoki Toda; | arxiv-cs.SD | 2022-11-14 |
280 | Exploiting Device and Audio Data to Tag Music with User-Aware Listening Contexts Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, for designing a fully automatised music retrieval system, we propose to disambiguate the user’s listening information from their stream data. |
Karim M. Ibrahim; Elena V. Epure; Geoffroy Peeters; Gaël Richard; | arxiv-cs.SD | 2022-11-14 |
281 | Pied Piper: Meta Search for Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose Pied piper, a meta search engine for music. |
Pulak Malhotra; Ashwin Rao; | arxiv-cs.IR | 2022-11-14 |
282 | Vis2Mus: Exploring Multimodal Representation Mapping for Controllable Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this study, we explore the representation mapping from the domain of visual arts to the domain of music, with which we can use visual arts as an effective handle to control music generation. |
Runbang Zhang; Yixiao Zhang; Kai Shao; Ying Shan; Gus Xia; | arxiv-cs.SD | 2022-11-10 |
283 | Museformer: Transformer with Fine- and Coarse-Grained Attention for Music Generation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose Museformer, a Transformer with a novel fine- and coarse-grained attention for symbolic music generation. |
BOTAO YU et. al. | nips | 2022-11-06 |
284 | Efficient Submodular Optimization Under Noise: Local Search Is Robust Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper designs a novel local search framework that can handle the effect of noise and achieve near-optimal approximation guarantees for submodular maximization with polynomial queries. |
Lingxiao Huang; Yuyi Wang; Chunxue Yang; Huanjian Zhou; | nips | 2022-11-06 |
285 | M4Singer: A Multi-Style, Multi-Singer and Musical Score Provided Mandarin Singing Corpus Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The lack of publicly available high-quality and accurately labeled datasets has long been a major bottleneck for singing voice synthesis (SVS). To tackle this problem, we present M4Singer, a free-to-use Multi-style, Multi-singer Mandarin singing collection with elaborately annotated Musical scores as well as its benchmarks. |
LICHAO ZHANG et. al. | nips | 2022-11-06 |
286 | Pop2Piano : Pop Audio-based Piano Cover Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present Pop2Piano, a Transformer network that generates piano covers given waveforms of pop music. |
Jongho Choi; Kyogu Lee; | arxiv-cs.SD | 2022-11-02 |
287 | Comparision Of Adversarial And Non-Adversarial LSTM Music Generative Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work implements and compares adversarial and non-adversarial training of recurrent neural network music composers on MIDI data. |
Moseli Mots’oehli; Anna Sergeevna Bosman; Johan Pieter De Villiers; | arxiv-cs.LG | 2022-11-01 |
288 | Self-Supervised Hierarchical Metrical Structure Modeling Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a novel method to model hierarchical metrical structures for both symbolic music and audio signals in a self-supervised manner with minimal domain knowledge. |
Junyan Jiang; Gus Xia; | arxiv-cs.SD | 2022-10-31 |
289 | SDMuse: Stochastic Differential Music Editing and Generation Via Hybrid Representation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose SDMuse, a unified Stochastic Differential Music editing and generation framework, which can not only compose a whole musical piece from scratch, but also modify existing musical pieces in many ways, such as combination, continuation, inpainting, and style transferring. |
Chen Zhang; Yi Ren; Kejun Zhang; Shuicheng Yan; | arxiv-cs.SD | 2022-10-31 |
290 | Analysis and Detection of Singing Techniques in Repertoires of J-POP Solo Singers Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we focus on singing techniques within the scope of music information retrieval research. |
Yuya Yamamoto; Juhan Nam; Hiroko Terasawa; | arxiv-cs.SD | 2022-10-31 |
291 | Relating Human Perception of Musicality to Prediction in A Predictive Coding Model Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: When trained with video data using self-supervised learning, the model manifests behaviors consistent with human visual illusions. Here, we adapt this network to model the hierarchical auditory system and investigate whether it will make similar choices to humans regarding the musicality of a set of random pitch sequences. |
NIKOLAS MCNEAL et. al. | arxiv-cs.SD | 2022-10-29 |
292 | LCD: Adaptive Label Correction for Denoising Music Recommendation Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Music recommendation is usually modeled as a Click-Through Rate (CTR) prediction problem, which estimates the probability of a user listening a recommended song. CTR prediction … |
QUANYU DAI et. al. | cikm | 2022-10-29 |
293 | An Empirical Study on How People Perceive AI-generated Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we evaluate human satisfaction with the state-of-the-art automatic symbolic music generation models using deep learning. |
HYESHIN CHU et. al. | cikm | 2022-10-29 |
294 | A Case Study in Educational Recommenders: Recommending Music Partitures at Tomplay Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Here we provide a study on the benefits of recommendation technologies in an educational platform with a focus on music learning. |
Ahmad Ajalloeian; Michalis Vlachos; Johannes Schneider; Alexis Steinmann; | cikm | 2022-10-29 |
295 | Music4All-Onion — A Large-Scale Multi-faceted Content-Centric Music Recommendation Dataset Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We organize distinct item content features in an onion model according to their semantics, and perform a comprehensive examination of the impact of different layers of this model (e.g., audio features, user-generated content, and derivative content) on content-driven music recommendation, demonstrating how various content features influence accuracy, novelty, and fairness of music recommendation systems. |
Marta Moscati; Emilia Parada-Cabaleiro; Yashar Deldjoo; Eva Zangerle; Markus Schedl; | cikm | 2022-10-29 |
296 | Discovery Dynamics: Leveraging Repeated Exposure for User and Music Characterization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we show that one particularly interesting repetitive behavior arises when users are consuming new items. |
Bruno Sguerra; Viet-Anh Tran; Romain Hennequin; | arxiv-cs.HC | 2022-10-28 |
297 | Spectrograms Are Sequences of Patches Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, only a tiny amount of work is related to music. In our work, we treat a spectrogram of music as a series of patches and design a self-supervised model that captures the features of these sequential patches: Patchifier, which makes good use of self-supervised learning methods from both NLP and CV domains. |
Leyi Zhao; Yi Li; | arxiv-cs.SD | 2022-10-28 |
298 | On The Role of Visual Context in Enriching Music Representations Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this study, we propose VCMR — Video-Conditioned Music Representations, a contrastive learning framework that learns music representations from audio and the accompanying music videos. |
Kleanthis Avramidis; Shanti Stewart; Shrikanth Narayanan; | arxiv-cs.SD | 2022-10-27 |
299 | Convolutive Block-Matching Segmentation Algorithm with Application to Music Structure Analysis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work presents a new algorithm, called Convolutive Block-Matching (CBM) algorithm, devoted to MSA. |
Axel Marmoret; Jérémy E. Cohen; Frédéric Bimbot; | arxiv-cs.SD | 2022-10-27 |
300 | LyricJam Sonic: A Generative System for Real-Time Composition and Musical Improvisation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Subsequently, it is difficult for artists to rediscover audio segments that might be suitable for use in their compositions from thousands of hours of recordings. In this paper, we describe LyricJam Sonic — a novel creative tool for musicians to rediscover their previous recordings, re-contextualize them with other recordings, and create original live music compositions in real-time. |
Olga Vechtomova; Gaurav Sahu; | arxiv-cs.SD | 2022-10-27 |
301 | A Survey on Artificial Intelligence for Music Generation: Agents, Domains and Perspectives Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Music generation with Artificial Intelligence is an emerging field that is gaining much attention in the recent years. In this paper, we describe how humans compose music and how new AI systems could imitate such process by comparing past and recent advances in the field with music composition techniques. |
Carlos Hernandez-Olivan; Javier Hernandez-Olivan; Jose R. Beltran; | arxiv-cs.AI | 2022-10-25 |
302 | Clarinet: A Music Retrieval System Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: A MIDI based approach for music recognition is proposed and implemented in this paper. |
Kshitij Alwadhi; Rohan Sharma; Siddhant Sharma; | arxiv-cs.IR | 2022-10-23 |
303 | A Computational Analysis on The Relationship Between Melodic Originality and Thematic Fame in Classical Music from The Romantic Period Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, the researcher presents a novel approach to calculating melodic originality based on the research by Simonton (1994). |
Hudson Griffith; | arxiv-cs.MM | 2022-10-21 |
304 | Data-Augmented Counterfactual Learning for Bundle Recommendation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel graph learning paradigm called Counterfactual Learning for Bundle Recommendation (CLBR) to mitigate the impact of data sparsity problem and improve bundle recommendation. |
Shixuan Zhu; Qi Shen; Yiming Zhang; Zhenwei Dong; Zhihua Wei; | arxiv-cs.IR | 2022-10-19 |
305 | Quantized GAN for Complex Music Generation from Dance Videos IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present Dance2Music-GAN (D2M-GAN), a novel adversarial multi-modal framework that generates complex musical samples conditioned on dance videos. |
YE ZHU et. al. | eccv | 2022-10-19 |
306 | Modelling Emotion Dynamics in Song Lyrics with State Space Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a method to predict emotion dynamics in song lyrics without song-level supervision. |
Yingjin Song; Daniel Beck; | arxiv-cs.CL | 2022-10-17 |
307 | DiffRoll: Diffusion-based Generative Music Transcription with Unsupervised Pretraining Capability Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we propose a novel generative approach, DiffRoll, to tackle automatic music transcription (AMT). |
KIN WAI CHEUK et. al. | arxiv-cs.SD | 2022-10-11 |
308 | ConchShell: A Generative Adversarial Networks That Turns Pictures Into Piano Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present ConchShell, a multi-modal generative adversarial framework that takes pictures as input to the network and generates piano music samples that match the picture context. |
Wanpeng Fan; Yuanzhi Su; Yuxin Huang; | arxiv-cs.SD | 2022-10-10 |
309 | Supervised and Unsupervised Learning of Audio Representations for Music Understanding IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we provide a broad comparative analysis of strategies for pre-training audio understanding models for several tasks in the music domain, including labelling of genre, era, origin, mood, instrumentation, key, pitch, vocal characteristics, tempo and sonority. |
Matthew C. McCallum; Filip Korzeniowski; Sergio Oramas; Fabien Gouyon; Andreas F. Ehmann; | arxiv-cs.SD | 2022-10-07 |
310 | AnimeTAB: A New Guitar Tablature Dataset of Anime and Game Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present AnimeTAB, a fingerstyle guitar tablature dataset in MusicXML format, which provides more high-quality guitar tablature for both researchers and guitar players. |
Yuecheng Zhou; Yaolong Ju; Lingyun Xie; | arxiv-cs.SD | 2022-10-06 |
311 | Melody Infilling with User-Provided Structural Context Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper proposes a novel Transformer-based model for music score infilling, to generate a music passage that fills in the gap between given past and future contexts. |
Chih-Pin Tan; Alvin W. Y. Su; Yi-Hsuan Yang; | arxiv-cs.SD | 2022-10-06 |
312 | Automatic Generation of Product Concepts from Positive Examples, with An Application to Music Streaming Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we represent product concepts using database queries and tackle two learning problems. |
KSHITIJ GOYAL et. al. | arxiv-cs.LG | 2022-10-04 |
313 | And What If Two Musical Versions Don’t Share Melody, Harmony, Rhythm, or Lyrics ? Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: On the other hand, using features focusing on specific aspects of musical pieces, such as melody, harmony, or lyrics, yielded interpretable and promising performances. In this work, we build upon these recent advances and propose a metric learning-based system systematically leveraging four dimensions commonly admitted to convey musical similarity between versions: melodic line, harmonic structure, rhythmic patterns, and lyrics. |
Mathilde Abrassart; Guillaume Doras; | arxiv-cs.SD | 2022-10-03 |
314 | EmoMV: Affective Music-video Correspondence Learning Datasets for Classification and Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts View |
Ha Thi Phuong Thao; Dorien Herremans; G. Roig; | Inf. Fusion | 2022-10-01 |
315 | Song Emotion Recognition: A Performance Comparison Between Audio Features and Artificial Neural Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Usually, it takes a data-set, from which audio features are extracted to present this information to a data-driven model, that will, in turn, train to predict what is the probability that a given song matches a target emotion. In this paper, we studied the most common features and models used in recent publications to tackle this problem, revealing which ones are best suited for recognizing emotion in a cappella songs. |
Karen Rosero; Arthur Nicholas dos Santos; Pedro Benevenuto Valadares; Bruno Sanches Masiero; | arxiv-cs.SD | 2022-09-24 |
316 | Maths, Computation and Flamenco: Overview and Challenges Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Over the recent decade, a number of computational problems related to the automatic analysis of flamenco music have been defined and several methods addressing a variety of musical aspects have been proposed. |
José-Miguel Díaz-Báñez; Nadine Kroher; | arxiv-cs.SD | 2022-09-22 |
317 | Learning Hierarchical Metrical Structure Beyond Measures Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we explore a data-driven approach to automatically extract hierarchical metrical structures from scores. |
Junyan Jiang; Daniel Chin; Yixiao Zhang; Gus Xia; | arxiv-cs.SD | 2022-09-21 |
318 | Setting The Rhythm Scene: Deep Learning-based Drum Loop Generation from Arbitrary Language Cues Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Here we present a novel method that, given an English word or phrase, generates 2 compasses of a 4-piece drum pattern that embodies the mood of the given language cue, or that could be used for an audiovisual scene described by the language cue. |
Ignacio J. Tripodi; | arxiv-cs.SD | 2022-09-20 |
319 | WideResNet with Joint Representation Learning and Data Augmentation for Cover Song Identification Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Cover song identification (CSI) has been a challenging task and an import topic in music information retrieval (MIR) commu-nity. In recent years, CSI problems have been … |
SHICHAO HU et. al. | Interspeech | 2022-09-18 |
320 | A Holistic Approach for Image-to-graph: Application to Optical Music Recognition Related Papers Related Patents Related Grants Related Venues Related Experts View |
Carlos Garrido-Munoz; Antonio Ríos-Vila; Jorge Calvo-Zaragoza; | International Journal on Document Analysis and Recognition … | 2022-09-16 |
321 | Musicaiz: A Python Library for Symbolic Music Generation, Analysis and Visualization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this article, we present musicaiz, an object-oriented library for analyzing, generating and evaluating symbolic music. |
Carlos Hernandez-Olivan; Jose R. Beltran; | arxiv-cs.SD | 2022-09-16 |
322 | Compose & Embellish: Well-Structured Piano Performance Generation Via A Two-Stage Approach Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Observing the above, we devise a two-stage Transformer-based framework that Composes a lead sheet first, and then Embellishes it with accompaniment and expressive touches. |
Shih-Lun Wu; Yi-Hsuan Yang; | arxiv-cs.SD | 2022-09-16 |
323 | Domain Adversarial Training on Conditional Variational Auto-Encoder for Controllable Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Specifically, we propose a condition corruption objective that uses the representation to denoise a corrupted condition. |
Jingwei Zhao; Gus Xia; Ye Wang; | arxiv-cs.SD | 2022-09-15 |
324 | MR4MR: Mixed Reality for Melody Reincarnation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we describe MR4MR, a sound installation work that allows users to experience melodies produced from interactions with their surrounding space in the context of Mixed Reality (MR). |
ATSUYA KOBAYASHI et. al. | arxiv-cs.HC | 2022-09-14 |
325 | A Decision-support System for Assessing The Function of Machine Learning and Artificial Intelligence in Music Education for Network Games Related Papers Related Patents Related Grants Related Venues Related Experts View |
Z. YUN et. al. | Soft Computing | 2022-09-13 |
326 | SongDriver: Real-time Music Accompaniment Generation Without Logical Latency Nor Exposure Bias Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose SongDriver, a real-time music accompaniment generation system without logical latency nor exposure bias. |
ZIHAO WANG et. al. | arxiv-cs.SD | 2022-09-13 |
327 | MATT: A Multiple-instance Attention Mechanism for Long-tail Music Genre Classification Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Inspired by the success of introducing Multi-instance Learning (MIL) in various classification tasks, we propose a novel mechanism named Multi-instance Attention (MATT) to boost the performance for identifying tail classes. |
Xiaokai Liu; Menghua Zhang; | arxiv-cs.SD | 2022-09-08 |
328 | Improving Choral Music Separation Through Expressive Synthesized Data from Sampled Instruments Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we investigate the use of synthesized training data for the source separation task on real choral music. |
KE CHEN et. al. | arxiv-cs.SD | 2022-09-06 |
329 | Bridging Music and Text with Crowdsourced Music Comments: A Sequence-to-Sequence Framework for Thematic Music Comments Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we exploit the crowd-sourced music comments to construct a new dataset and propose a sequence-to-sequence model to generate text descriptions of music. |
Peining Zhang; Junliang Guo; Linli Xu; Mu You; Junming Yin; | arxiv-cs.SD | 2022-09-05 |
330 | More Than Words: Linking Music Preferences and Moral Values Through Lyrics Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This study explores the association between music preferences and moral values by applying text analysis techniques to lyrics. |
Vjosa Preniqi; Kyriaki Kalimeri; Charalampos Saitis; | arxiv-cs.CY | 2022-09-02 |
331 | Dynamic Session-based Music Recommendation Using Information Retrieval Techniques Related Papers Related Patents Related Grants Related Venues Related Experts View |
Arthur Tofani; Rodrigo Borges; Marcelo Queiroz; | User Modeling and User-Adapted Interaction | 2022-09-01 |
332 | AccoMontage2: A Complete Harmonization and Accompaniment Arrangement System Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose AccoMontage2, a system capable of doing full-length song harmonization and accompaniment arrangement based on a lead melody. |
Li Yi; Haochen Hu; Jingwei Zhao; Gus Xia; | arxiv-cs.SD | 2022-09-01 |
333 | Generating Coherent Drum Accompaniment With Fills And Improvisations Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we tackle the task of drum pattern generation conditioned on the accompanying music played by four melodic instruments: Piano, Guitar, Bass, and Strings. |
Rishabh Dahale; Vaibhav Talwadker; Preeti Rao; Prateek Verma; | arxiv-cs.SD | 2022-09-01 |
334 | A Musical Serious Game for Social Interaction Through Augmented Rhythmic Improvisation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a serious game that helps people even without musical skills interact with each other by collaboratively creating a rhythm with MIDI drum pads. |
Filippo Carnovalini; Antonio Rodà; Paolo Caneva; | arxiv-cs.MM | 2022-08-31 |
335 | What Is Missing in Deep Music Generation? A Study of Repetition and Structure in Popular Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Analyses of two popular music datasets (Chinese and American) illustrate important music construction principles: (1) structure exists at multiple hierarchical levels, (2) songs use repetition and limited vocabulary so that individual songs do not follow general statistics of song collections, (3) structure interacts with rhythm, melody, harmony, and predictability, and (4) over the course of a song, repetition is not random, but follows a general trend as revealed by cross-entropy. These and other findings offer challenges as well as opportunities for deep-learning music generation and suggest new formal music criteria and evaluation methods. |
Shuqi Dai; Huiran Yu; Roger B. Dannenberg; | arxiv-cs.SD | 2022-08-31 |
336 | MeloForm: Generating Melody with Musical Form Based on Expert Systems and Neural Networks Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we develop MeloForm, a system that generates melody with musical form using expert systems and neural networks. |
PEILING LU et. al. | arxiv-cs.SD | 2022-08-30 |
337 | Towards Robust Music Source Separation on Loud Commercial Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Yet, in music source separation, these characteristics have not been thoroughly considered, resulting in the domain mismatch between the laboratory and the real world. In this paper, we confirmed that this domain mismatch negatively affect the performance of the music source separation networks. |
Chang-Bin Jeon; Kyogu Lee; | arxiv-cs.SD | 2022-08-30 |
338 | Multi-objective Hyper-parameter Optimization of Behavioral Song Embeddings Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we study the hyper-parameter optimization of behavioral song embeddings based on Word2Vec on a selection of downstream tasks, namely next-song recommendation, false neighbor rejection, and artist and genre clustering. |
Massimo Quadrana; Antoine Larreche-Mouly; Matthias Mauch; | arxiv-cs.IR | 2022-08-26 |
339 | Music Separation Enhancement with Generative Modeling Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a post-processing model (the Make it Sound Good (MSG) post-processor) to enhance the output of music source separation systems. |
NOAH SCHAFFER et. al. | arxiv-cs.SD | 2022-08-25 |
340 | A Study on Broadcast Networks for Music Genre Classification Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we study the broadcast-based neural networks aiming to improve the localization and generalizability under a small set of parameters (about 180k) and investigate twelve variants of broadcast networks discussing the effect of block configuration, pooling method, activation function, normalization mechanism, label smoothing, channel interdependency, LSTM block inclusion, and variants of inception schemes. |
Ahmed Heakl; Abdelrahman Abdelgawad; Victor Parque; | arxiv-cs.SD | 2022-08-25 |
341 | Contrastive Audio-Language Learning for Music IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we explore cross-modal learning in an attempt to bridge audio and language in the music domain. |
Ilaria Manco; Emmanouil Benetos; Elio Quinton; György Fazekas; | arxiv-cs.SD | 2022-08-25 |
342 | Interpreting Song Lyrics with An Audio-Informed Pre-trained Language Model Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper we propose BART-fusion, a novel model for generating lyric interpretations from lyrics and music audio that combines a large-scale pre-trained language model with an audio encoder. |
Yixiao Zhang; Junyan Jiang; Gus Xia; Simon Dixon; | arxiv-cs.SD | 2022-08-24 |
343 | Exploring Popularity Bias in Music Recommendation Models and Commercial Steaming Services Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we attempt to measure popularity bias in three state-of-art recommender system models (e.g., SLIM, Multi-VAE, WRMF) and on three commercial music streaming services (Spotify, Amazon Music, YouTube). |
Douglas R. Turnbull; Sean McQuillan; Vera Crabtree; John Hunter; Sunny Zhang; | arxiv-cs.IR | 2022-08-19 |
344 | Melatonin: A Case Study on AI-induced Musical Style Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we present a case study of Melatonin, a song produced by extensive use of BassNet, an AI tool originally designed to generate bass lines. |
Emmanuel Deruty; Maarten Grachten; | arxiv-cs.AI | 2022-08-18 |
345 | Musika! Fast Infinite Waveform Music Generation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we introduce Musika, a music generation system that can be trained on hundreds of hours of music using a single consumer GPU, and that allows for much faster than real-time generation of music of arbitrary length on a consumer CPU. |
Marco Pasini; Jan Schlüter; | arxiv-cs.SD | 2022-08-18 |
346 | Extract Fundamental Frequency Based on CNN Combined with PYIN Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper refers to the extraction of multiple fundamental frequencies (multiple F0) based on PYIN, an algorithm for extracting the fundamental frequency (F0) of monophonic music, and a trained convolutional neural networks (CNN) model, where a pitch salience function of the input signal is produced to estimate the multiple F0. The implementation of these two algorithms and their corresponding advantages and disadvantages are discussed in this article. |
Ruowei Xing; Shengchen Li; | arxiv-cs.SD | 2022-08-17 |
347 | A User-Centered Investigation of Personal Music Tours Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we provide, for the first time, a user-centered evaluation of two tour-generation algorithms (Greedy and Optimal) using semi-structured interviews. |
Giovanni Gabbolini; Derek Bridge; | arxiv-cs.IR | 2022-08-16 |
348 | Models of Music Cognition and Composition Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Much like most of cognition research, music cognition is an interdisciplinary field, which attempts to apply methods of cognitive science (neurological, computational and experimental) to understand the perception and process of composition of music. |
Abhimanyu Sethia; | arxiv-cs.SD | 2022-08-14 |
349 | Neural Encoding of Songs Is Modulated By Their Enjoyment Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We examine user and song identification from neural (EEG) signals. |
Gulshan Sharma; Pankaj Pandey; Ramanathan Subramanian; Krishna. P. Miyapuram; Abhinav Dhall; | arxiv-cs.HC | 2022-08-13 |
350 | Re-creation of Creations: A New Paradigm for Lyric-to-Melody Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose Re-creation of Creations (ROC), a new paradigm for lyric-to-melody generation. |
Ang Lv; Xu Tan; Tao Qin; Tie-Yan Liu; Rui Yan; | arxiv-cs.SD | 2022-08-11 |
351 | Controlling Perceived Emotion in Symbolic Music Generation with Monte Carlo Tree Search Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper presents a new approach for controlling emotion in symbolic music generation with Monte Carlo Tree Search. |
Lucas N. Ferreira; Lili Mou; Jim Whitehead; Levi H. S. Lelis; | arxiv-cs.SD | 2022-08-10 |
352 | Symbolic Music Loop Generation with Neural Discrete Representations Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose symbolic music loop generation via learning discrete representations. |
Sangjun Han; Hyeongrae Ihm; Moontae Lee; Woohyung Lim; | arxiv-cs.SD | 2022-08-10 |
353 | Pure Data and INScore: Animated Notation for New Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This essay will show a case study that uses Pure Data, in connection with INScore, for making a new type of score that uses animated notation or dynamic musicography for making music with performers. |
Patricio F. Calatayud; | arxiv-cs.SD | 2022-08-09 |
354 | Debiased Cross-modal Matching for Content-based Micro-video Background Music Recommendation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, the selection of the user-generated content (UGC) is biased caused by knowledge limitations and historical preferences among music of each uploader. In this paper, we propose a Debiased Cross-Modal (DebCM) matching model to alleviate the influence of such selection bias. |
Jinng Yi; Zhenzhong Chen; | arxiv-cs.MM | 2022-08-07 |
355 | A Sequential MUSIC Algorithm for Scatterers Detection 2 in SAR Tomography Enhanced By A Robust Covariance 3 Estimator Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, a new sequential MUSIC algorithm named recursive covariance canceled MUSIC (RCC-MUSIC), is proposed. |
Ahmad Naghavi; Mohammad Sadegh Fazel; Mojtaba Beheshti; Ehsan Yazdian; | arxiv-cs.IT | 2022-08-04 |
356 | Large Scale Analysis of Gender Bias and Sexism in Song Lyrics Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We employ Natural Language Processing techniques to analyse 377808 English song lyrics from the Two Million Song Database corpus, focusing on the expression of sexism across five decades (1960-2010) and the measurement of gender biases. |
Lorenzo Betti; Carlo Abrate; Andreas Kaltenbrunner; | arxiv-cs.CY | 2022-08-03 |
357 | Jazz Contrafact Detection Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper develops a novel vector-space model to represent chord progressions, and uses it for contrafact detection. |
C. Bunks; T. Weyde; | arxiv-cs.SD | 2022-08-01 |
358 | SampleMatch: Drum Sample Retrieval By Musical Context Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we explore the automatic drum sample retrieval based on aesthetic principles learned from data. |
Stefan Lattner; | arxiv-cs.SD | 2022-08-01 |
359 | Holistic Approaches to Music Genre Classification Using Efficient Transfer and Deep Learning Techniques Related Papers Related Patents Related Grants Related Venues Related Experts View |
S. Prabhakar; Seong-Whan Lee; | Expert Syst. Appl. | 2022-08-01 |
360 | Exploiting Negative Preference in Content-based Music Recommendation with Contrastive Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we analyze the role of negative preference in users’ music tastes by comparing music recommendation models with contrastive learning exploiting preference (CLEP) but with three different training strategies – exploiting preferences of both positive and negative (CLEP-PN), positive only (CLEP-P), and negative only (CLEP-N). |
Minju Park; Kyogu Lee; | arxiv-cs.IR | 2022-07-28 |
361 | EEG2Mel: Reconstructing Sound from Brain Responses to Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Information retrieval in the form of reconstructing auditory stimuli has also shown some success, but here we improve on previous methods by reconstructing music stimuli well enough to be perceived and identified independently. |
Adolfo G. Ramirez-Aristizabal; Chris Kello; | arxiv-cs.SD | 2022-07-27 |
362 | Unique in What Sense? Heterogeneous Relationships Between Multiple Types of Uniqueness and Popularity in Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper uses songs as a case study to study the multiple facets of uniqueness and their relationship with success. |
Yulin Yu; Pui Yin Cheung; Yong-Yeol Ahn; Paramveer Dhillon; | arxiv-cs.CY | 2022-07-26 |
363 | Perception-Aware Attack: Creating Adversarial Music Via Reverse-Engineering Human Perception Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we formulate the adversarial attack against music signals as a new perception-aware attack framework, which integrates human study into adversarial attack design. |
RUI DUAN et. al. | arxiv-cs.SD | 2022-07-26 |
364 | HouseX: A Fine-grained House Music Dataset and Its Potential in The Music Industry Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In consideration of the consistency of genres of songs in a mixtape or in a DJ (live) set, we have collected and annotated a dataset of house music that provide 4 sub-genre labels, namely future house, bass house, progressive house and melodic house. |
Xinyu Li; | arxiv-cs.SD | 2022-07-24 |
365 | The WASABI Song Corpus and Knowledge Graph for Music Lyrics Analysis Related Papers Related Patents Related Grants Related Venues Related Experts View |
MICHAEL FELL et. al. | Language Resources and Evaluation | 2022-07-21 |
366 | Multimodal Music Emotion Recognition with Hierarchical Cross-Modal Attention Network Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Computational music emotion recognition is to recognize the emotional content in music tracks. In computational music emotion recognition studies, researchers have paid close … |
JIAHAO ZHAO et. al. | 2022 IEEE International Conference on Multimedia and Expo … | 2022-07-18 |
367 | ChoreoGraph: Music-conditioned Automatic Dance Choreography Over A Style and Tempo Consistent Dynamic Graph Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: And finally, basic choreomusical rules should be observed, and the motion generated should be diverse. To address these challenges, we propose ChoreoGraph, which choreographs high-quality dance motion for a given piece of music over a Dynamic Graph. |
Ho Yin Au; Jie Chen; Junkun Jiang; Yike Guo; | arxiv-cs.MM | 2022-07-15 |
368 | Flow Moods: Recommending Music By Moods on Deezer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The music streaming service Deezer extensively relies on its Flow algorithm, which generates personalized radio-style playlists of songs, to help users discover musical content. … |
TH’EO BONTEMPELLI et. al. | Proceedings of the 16th ACM Conference on Recommender … | 2022-07-15 |
369 | Unaligned Supervision for Automatic Music Transcription in The Wild Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce Note$_{EM}$, a method for simultaneously training a transcriber and aligning the scores to their corresponding performances, in a fully-automated process. |
Ben Maman; Amit H Bermano; | icml | 2022-07-15 |
370 | Multitrack Music Transformer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a new multitrack music representation that allows a diverse set of instruments while keeping a short sequence length. |
Hao-Wen Dong; Ke Chen; Shlomo Dubnov; Julian McAuley; Taylor Berg-Kirkpatrick; | arxiv-cs.SD | 2022-07-14 |
371 | ReLyMe: Improving Lyric-to-Melody Generation By Incorporating Lyric-Melody Relationships Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose ReLyMe, a method that incorporates Relationships between Lyrics and Melodies from music theory to ensure the harmony between lyrics and melodies. |
CHEN ZHANG et. al. | arxiv-cs.SD | 2022-07-12 |
372 | Online Game Level Generation from Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose and formulate online level generation from music, in a way of matching a level feature to a music feature in real-time, while adapting to players’ play speed. |
Ziqi Wang; Jialin Liu; | arxiv-cs.AI | 2022-07-11 |
373 | Towards Proper Contrastive Self-supervised Learning Strategies For Music Audio Representation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we investigate music audio representation learned from different contrastive self-supervised learning schemes and empirically evaluate the embedded vectors on various music information retrieval (MIR) tasks where different levels of the music perception are concerned. |
Jeong Choi; Seongwon Jang; Hyunsouk Cho; Sehee Chung; | arxiv-cs.SD | 2022-07-10 |
374 | Music-driven Dance Regeneration with Controllable Key Pose Constraints Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel framework for music-driven dance motion synthesis with controllable key pose constraint. |
Junfu Pu; Ying Shan; | arxiv-cs.CV | 2022-07-08 |
375 | Learning Music-Dance Representations Through Explicit-Implicit Rhythm Synchronization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Considering the intrinsic alignment between the cadent movement of dancer and music rhythm, we introduce MuDaR, a novel Music-Dance Representation learning framework to perform the synchronization of music and dance rhythms both in explicit and implicit ways. |
Jiashuo Yu; Junfu Pu; Ying Cheng; Rui Feng; Ying Shan; | arxiv-cs.SD | 2022-07-07 |
376 | Music-CRN: An Efficient Content-Based Music Classification and Recommendation Network Related Papers Related Patents Related Grants Related Venues Related Experts View |
Yuxu Mao; Guoqiang Zhong; Haizhen Wang; Kaizhu Huang; | Cognitive Computation | 2022-07-05 |
377 | An Adaptive Music Generation Architecture for Games Based on The Deep Learning Transformer Mode Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents an architecture for generating music for video games based on the Transformer deep learning model. |
Gustavo Amaral Costa dos Santos; Augusto Baffa; Jean-Pierre Briot; Bruno Feijó; Antonio Luz Furtado; | arxiv-cs.SD | 2022-07-04 |
378 | Information Technology-based Revolution in Music Education Using AHP and TOPSIS Related Papers Related Patents Related Grants Related Venues Related Experts View |
Yi Fu; Mengjia Zhang; Muhammad Nawaz; Muhammad Ali; Aman Singh; | Soft Computing | 2022-07-04 |
379 | Music-to-Dance Generation with Optimal Transport Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a Music-to-Dance with Optimal Transport Network (MDOT-Net) for learning to generate 3D dance choreographies from music. |
Shuang Wu; Shijian Lu; Li Cheng; | ijcai | 2022-07-01 |
380 | Tradformer: A Transformer Model of Traditional Music Transcriptions Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We explore the transformer neural network architecture for modeling music, specifically Irish and Swedish traditional dance music. |
Luca Casini; Bob L. T. Sturm; | ijcai | 2022-07-01 |
381 | Multimodal Representation Learning Over Heterogeneous Networks for Tag-based Music Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts View |
A. Silva; Diego Furtado Silva; R. Marcacini; | Expert Syst. Appl. | 2022-07-01 |
382 | Interpretable Melody Generation from Lyrics with Discrete-Valued Adversarial Training Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In our proposal, we demonstrate our proposed interpretable lyrics-to-melody generation system which can interact with users to understand the generation process and recreate the desired songs. |
Wei Duan; Zhe Zhang; Yi Yu; Keizo Oyama; | arxiv-cs.SD | 2022-06-30 |
383 | Music-to-Dance Generation with Multiple Conformer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: It is necessary for the music-to-dance generation to consider both the kinematics in dance that is highly complex and non-linear and the connection between music and dance … |
Mingao Zhang; Changhong Liu; Yong Chen; Zhenchun Lei; Mingwen Wang; | Proceedings of the 2022 International Conference on … | 2022-06-27 |
384 | Calliope: A Co-creative Interface for Multi-Track Music Generation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Calliope is a web application for co-creative multi-track music composition (MMM) in the symbolic domain. It is built to facilitate the use of multi-track music machine (MMM). The … |
R. Tchemeube; Jeffrey John Ens; P. Pasquier; | Proceedings of the 14th Conference on Creativity and … | 2022-06-20 |
385 | Novelty and Cultural Evolution in Modern Popular Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In order to uncover underlying patterns in cultural evolution, we examine the relationship between the established characteristics of different genres and styles, and the introduction of novel ideas that fuel this ongoing creative evolution. To understand how this dynamic plays out and shapes the cultural ecosystem, we compare musical artifacts to their contemporaries to identify novel artifacts, study the relationship between novelty and commercial success, and connect this to the changes in musical content that we can observe over time. |
Katherine O’Toole; Emőke-Ágnes Horvát; | arxiv-cs.CY | 2022-06-15 |
386 | It’s Time for Artistic Correspondence in Music and Video IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present an approach for recommending a music track for a given video, and vice versa, based on both their temporal alignment and their correspondence at an artistic level. |
Didac Suris; Carl Vondrick; Bryan Russell; Justin Salamon; | arxiv-cs.MM | 2022-06-14 |
387 | Self-Supervised Representation Learning With MUlti-Segmental Informational Coding (MUSIC) Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we propose MUlti-Segmental Informational Coding (MUSIC) for self-supervised representation learning. |
Chuang Niu; Ge Wang; | arxiv-cs.CV | 2022-06-13 |
388 | Bailando: 3D Dance Generation By Actor-Critic GPT With Choreographic Memory IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In addition, the generated dance sequence also needs to maintain temporal coherency with different music genres. To tackle these challenges, we propose a novel music-to-dance framework, Bailando, with two powerful components: 1) a choreographic memory that learns to summarize meaningful dancing units from 3D pose sequence to a quantized codebook, 2) an actor-critic Generative Pre-trained Transformer (GPT) that composes these units to a fluent dance coherent to the music. |
LI SIYAO et. al. | cvpr | 2022-06-07 |
389 | It’s Time for Artistic Correspondence in Music and Video Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present an approach for recommending a music track for a given video, and vice versa, based on both their temporal alignment and their correspondence at an artistic level. |
Dídac Surís; Carl Vondrick; Bryan Russell; Justin Salamon; | cvpr | 2022-06-07 |
390 | The Musical Arrow of Time — The Role of Temporal Asymmetry in Music and Its Organicist Implications Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Adopting a performer-centric perspective, we frequently encounter two statements: music flows, and music is life-like. |
Qi Xu; | arxiv-cs.SD | 2022-06-02 |
391 | Partitura: A Python Package for Symbolic Music Processing Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: Partitura is a lightweight Python package for handling symbolic musical information. It provides easy access to features commonly used in music information retrieval tasks, like … |
CARLOS CANCINO-CHACÓN et. al. | arxiv-cs.SD | 2022-06-02 |
392 | Co-creation and Ownership for AI Radio Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present Artificial.fm, a proof-of-concept casual creator that blends AI-music generation, subjective ratings, and personalized recommendation for the creation and curation of AI-generated music. |
Skylar Gordon; Robert Mahari; Manaswi Mishra; Ziv Epstein; | arxiv-cs.CY | 2022-06-01 |
393 | The Hipster Paradox in Electronic Dance Music: How Musicians Trade Mainstream Success Off Against Alternative Status Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We construct network snapshots following a formal sociological approach based on bipartite networks, and we use network positions to explain success in regression models of artistic careers. |
Mohsen Jadidi; Haiko Lietz; Mattia Samory; Claudia Wagner; | arxiv-cs.SI | 2022-06-01 |
394 | Self-Supervised Music Motion Synchronization Learning for Music-Driven Conducting Motion Generation Related Papers Related Patents Related Grants Related Venues Related Experts View |
Fan Liu; Delong Chen; Rui-Zhi Zhou; Sai Yang; Feng Xu; | Journal of Computer Science and Technology | 2022-05-31 |
395 | Singer Identification for Metaverse with Timbral and Middle-Level Perceptual Features IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: It means the music information is noise for using frame-level features to identify the singers. In this paper, instead of only the frame-level features, we propose to use another two features that address this problem. |
Xulong Zhang; Jianzong Wang; Ning Cheng; Jing Xiao; | arxiv-cs.SD | 2022-05-24 |
396 | MetaSID: Singer Identification with Domain Adaptation for Metaverse Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes the use of the domain adaptation method to solve the live effect in singer identification. |
Xulong Zhang; Jianzong Wang; Ning Cheng; Jing Xiao; | arxiv-cs.SD | 2022-05-24 |
397 | SUSing: SU-net for Singing Voice Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we proposed SU-net for singing voice synthesis named SUSing. |
Xulong Zhang; Jianzong Wang; Ning Cheng; Jing Xiao; | arxiv-cs.SD | 2022-05-24 |
398 | Source Separation By Steering Pretrained Music Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We use OpenAI�s Jukebox as the pretrained generative model, and we couple it with four kinds of pretrained music taggers (two architectures and two tagging datasets). |
E. Manilow; P. O�Reilly; P. Seetharaman; B. Pardo; | icassp | 2022-05-22 |
399 | Learning Music Audio Representations Via Weak Language Supervision IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we pose the question of whether it may be possible to exploit weakly aligned text as the only supervisory signal to learn general-purpose music audio representations. |
I. Manco; E. Benetos; E. Quinton; G. Fazekas; | icassp | 2022-05-22 |
400 | To Catch A Chorus, Verse, Intro, or Anything Else: Analyzing A Song with Structural Functions IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, explicitly identifying the function of each segment (e.g., �verse� or �chorus�) is rarely attempted, but has many applications. We introduce a multi-task deep learning framework to model these structural semantic labels directly from audio by estimating verseness, chorusness, and so forth, as a function of time. |
J. -C. Wang; Y. -N. Hung; J. B. L. Smith; | icassp | 2022-05-22 |
401 | A Novel 1D State Space for Efficient Music Rhythmic Analysis Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper proposes a new state space and a semi-Markov model for music time structure analysis. |
M. Heydari; M. McCallum; A. Ehmann; Z. Duan; | icassp | 2022-05-22 |
402 | Upmixing Via Style Transfer: A Variational Autoencoder for Disentangling Spatial Images And Musical Content Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a modified variational autoencoder model that learns a latent space to describe the spatial images in multichannel music. |
H. YANG et. al. | icassp | 2022-05-22 |
403 | SleepGAN: Towards Personalized Sleep Therapy Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we take the first step towards generating personalized sleep therapy music. |
J. Yang; C. Min; A. Mathur; F. Kawsar; | icassp | 2022-05-22 |
404 | Tonet: Tone-Octave Network for Singing Melody Extraction from Polyphonic Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose TONet1, a plug-and-play model that improves both tone and octave perceptions by leveraging a novel input representation and a novel network architecture. |
K. CHEN et. al. | icassp | 2022-05-22 |
405 | Ambiguity Modelling with Label Distribution Learning for Music Classification Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we address the issue of ambiguity that can arise in many classification problems. |
M. Buisson; P. Alonso-Jim�nez; D. Bogdanov; | icassp | 2022-05-22 |
406 | End-To-End Music Remastering System Using Self-Supervised And Adversarial Training Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Remastering follows the same technical process, in which the context lies in mastering a song for the times. As these tasks have high entry barriers, we aim to lower the barriers by proposing an end-to-end music remastering system that transforms the mastering style of input audio to that of the target. |
J. Koo; S. Paik; K. Lee; | icassp | 2022-05-22 |
407 | Automatic DJ Transitions with Differentiable Audio Effects and Generative Adversarial Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we explore a data-driven approach that uses a generative adversarial network to create the song transition by learning from real-world DJ mixes. |
B. -Y. CHEN et. al. | icassp | 2022-05-22 |
408 | Training Strategies for Automatic Song Writing: A Unified Framework Perspective Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a unified framework following the pre-training and fine-tuning paradigm to address all four ASW tasks with one model. |
T. Qian; J. Shi; S. Guo; P. Wu; Q. Jin; | icassp | 2022-05-22 |
409 | Learning Music Sequence Representation From Text Supervision Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To excavate better MUsic SEquence Representation from labeled audio, we propose a novel text-supervision pre-training method, namely MUSER. |
T. CHEN et. al. | icassp | 2022-05-22 |
410 | Genre-Conditioned Long-Term 3D Dance Generation Driven By Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we focus on generating long-term 3D dance from music with a specific genre. |
Y. Huang; et al. | icassp | 2022-05-22 |
411 | Hierarchical Graph-Based Neural Network for Singing Melody Extraction Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel hierarchical graph-based network for singing melody extraction. |
S. Yu; X. Chen; W. Li; | icassp | 2022-05-22 |
412 | Bytecover2: Towards Dimensionality Reduction of Latent Embedding for Efficient Cover Song Identification IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose an up-graded version of ByteCover, termed ByteCover2, which further improves ByteCover in both identification performance and efficiency. |
X. Du; K. Chen; Z. Wang; B. Zhu; Z. Ma; | icassp | 2022-05-22 |
413 | A Melody-Unsupervision Model for Singing Voice Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To address the issue, we propose a melody-unsupervision model that requires only audio-and-lyrics pairs without temporal alignment in training time but generates singing voice audio given a melody and lyrics input in inference time. |
S. Choi; J. Nam; | icassp | 2022-05-22 |
414 | Hierarchical Classification of Singing Activity, Gender, and Type in Complex Music Recordings Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Traditionally, work on singing voice detection has focused on identifying singing activity in music recordings. In this work, our aim is to extend this task towards simultaneously detecting the presence of singing voice as well as determining singer gender and voice type. |
M. Krause; M. M�ller; | icassp | 2022-05-22 |
415 | Genre-Conditioned Acoustic Models for Automatic Lyrics Transcription of Polyphonic Music IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose to transcribe the lyrics of polyphonic music using a novel genre-conditioned network. |
X. Gao; C. Gupta; H. Li; | icassp | 2022-05-22 |
416 | Deep Augmented Music Algorithm for Data-Driven Doa Estimation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper introduces a new hybrid MB/DD DoA estimation architecture, based on the classical multiple signal classification (MUSIC) algorithm. |
J. P. Merkofer; G. Revach; N. Shlezinger; R. J. G. van Sloun; | icassp | 2022-05-22 |
417 | S3T: Self-Supervised Pre-Training with Swin Transformer For Music Classification Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose S3T, a self-supervised pre-training method with Swin Transformer for music classification, aiming to learn meaningful music representations from massive easily accessible unlabeled music data. |
H. Zhao; C. Zhang; B. Zhu; Z. Ma; K. Zhang; | icassp | 2022-05-22 |
418 | Music Phrase Inpainting Using Long-Term Representation and Contrastive Loss Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we tackle the problem of long-term, phrase-level symbolic melody inpainting by equipping a sequence prediction model with phrase-level representation (as an extra condition) and contrastive loss (as an extra optimization term). |
S. Wei; G. Xia; Y. Zhang; L. Lin; W. Gao; | icassp | 2022-05-22 |
419 | Music Enhancement Via Image Translation and Vocoding Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Consumer-grade music recordings such as those captured by mobile devices typically contain distortions in the form of background noise, reverb, and microphone-induced EQ. This paper presents a deep learning approach to enhance low-quality music recordings by combining (i) an image-to-image translation model for manipulating audio in its mel-spectrogram representation and (ii) a music vocoding model for mapping synthetically generated mel-spectrograms to perceptually realistic waveforms. |
N. Kandpal; O. Nieto; Z. Jin; | icassp | 2022-05-22 |
420 | Music Identification Using Brain Responses to Initial Snippets Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We examine EEG encoding of naturalistic musical patterns employing the NMED-T and MUSIN-G datasets. |
P. Pandey; G. Sharma; K. P. Miyapuram; R. Subramanian; D. Lomas; | icassp | 2022-05-22 |
421 | Melons: Generating Melody With Long-Term Structure Using Transformers And Structure Graph IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose MELONS, a melody generation framework based on a graph representation of music structure which consists of eight types of bar-level relations. |
Y. ZOU et. al. | icassp | 2022-05-22 |
422 | Deep Performer: Score-to-Audio Music Performance Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Hence, we propose two new techniques for handling polyphonic inputs and providing a fine-grained conditioning in a transformer encoder-decoder model. |
H. -W. Dong; C. Zhou; T. Berg-Kirkpatrick; J. McAuley; | icassp | 2022-05-22 |
423 | Enhancing Affective Representations Of Music-Induced Eeg Through Multimodal Supervision And Latent Domain Adaptation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper we extract efficient, personalized affective representations from EEG signals during music listening. |
K. Avramidis; C. Garoufis; A. Zlatintsi; P. Maragos; | icassp | 2022-05-22 |
424 | Audio-To-Symbolic Arrangement Via Cross-Modal Music Representation Learning Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Could we automatically derive the score of a piano accompaniment based on the audio of a pop song? This is the audio-to-symbolic arrangement problem we tackle in this paper. |
Z. Wang; D. Xu; G. Xia; Y. Shan; | icassp | 2022-05-22 |
425 | Progressive Teacher-Student Training Framework for Music Tagging Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a progressive two-stage teacher-student training framework to prevent the music tagging model from overfitting label noise. |
R. LU et. al. | icassp | 2022-05-22 |
426 | An Automated System Recommending Background Music to Listen to While Working Related Papers Related Patents Related Grants Related Venues Related Experts View |
Hiromu Yakura; Tomoyasu Nakano; Masataka Goto; | User Modeling and User-Adapted Interaction | 2022-05-18 |
427 | The Power of Fragmentation: A Hierarchical Transformer Model for Structural Segmentation in Symbolic Music Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a hierarchical Transformer model to learn multi-scale contexts in music. |
Guowei Wu; Shipei Liu; Xiaoya Fan; | arxiv-cs.SD | 2022-05-17 |
428 | Transformer-based Ensemble Method for Multiple Predominant Instruments Recognition in Polyphonic Music Related Papers Related Patents Related Grants Related Venues Related Experts View |
L. Reghunath; R. Rajan; | EURASIP Journal on Audio, Speech, and Music Processing | 2022-05-16 |
429 | Conditional Vector Graphics Generation for Music Cover Images Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, they belong to the highly sophisticated parametric space, which is a restriction for solving the task of synthesizing vector graphics by GANs. In this paper, we consider a specific application domain that softens this restriction dramatically allowing the usage of vector image synthesis. |
Valeria Efimova; Ivan Jarsky; Ilya Bizyaev; Andrey Filchenkov; | arxiv-cs.GR | 2022-05-15 |
430 | CMelGAN: An Efficient Conditional Generative Model Based on Mel Spectrograms Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: There are many applications of machine learning in music, like the classifying the mood of a piece of music, conditional music generation, or popularity prediction. The goal for this project was to develop a genre-conditional generative model of music based on Mel spectrograms and evaluate its performance by comparing it to existing generative music models that use note-based representations. |
Tracy Qian; Jackson Kaunismaa; Tony Chung; | arxiv-cs.SD | 2022-05-15 |
431 | Generating Tips from Song Reviews: A New Dataset and Framework Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we create a dataset named MTips for the task and propose a framework named GENTMS for automatically generating tips from song reviews. |
JINGYA ZANG et. al. | arxiv-cs.IR | 2022-05-14 |
432 | Bangla Music Lyrics Classification Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Music is one of the most important factors of human lives. People express their inner thoughts, emotions and feelings with the combination of both lyrics and musical instruments. … |
Shafi Ahmed; Md Humaion Kabir Mehedi; Moh. Absar Rahman; Jawad Bin Sayed; | Proceedings of the 2022 8th International Conference on … | 2022-05-12 |
433 | An Adaptive Meta-heuristic for Music Plagiarism Detection Based on Text Similarity and Clustering Related Papers Related Patents Related Grants Related Venues Related Experts View |
Delfina Malandrino; R. De Prisco; Mario Ianulardo; R. Zaccagnino; | Data Mining and Knowledge Discovery | 2022-05-12 |
434 | Scream Detection in Heavy Metal Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The main contributions of this work are (i) a manually annotated dataset comprised of over 280 minutes of heavy metal songs of various genres with a statistical analysis of occurrences of different extreme vocal techniques in heavy metal music, and (ii) a systematic study of different input feature representations for the classification of heavy metal vocals |
Vedant Kalbag; Alexander Lerch; | arxiv-cs.SD | 2022-05-11 |
435 | Symphony Generation with Permutation Invariant Language Model IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a permutation invariant language model, SymphonyNet, as a solution for symbolic symphony music generation. |
JIAFENG LIU et. al. | arxiv-cs.SD | 2022-05-10 |
436 | Muskits: An End-to-End Music Processing Toolkit for Singing Voice Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper introduces a new open-source platform named Muskits for end-to-end music processing, which mainly focuses on end-to-end singing voice synthesis (E2E-SVS). |
JIATONG SHI et. al. | arxiv-cs.SD | 2022-05-09 |
437 | A Music-Therapy Robotic Platform for Children with Autism: A Pilot Study Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a novel robot-based music-therapy platform for modeling and improving the social responses and behaviors of children with ASD. |
Huanghao Fengr; Mohammad H. Mahoor; Francesca Dino; | arxiv-cs.HC | 2022-05-09 |
438 | Psychologically-Inspired Music Recommendation System Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Our goal is to find a way to integrate users’ personal traits and their current emotional state into a single music recommendation system with both collaborative and content-based filtering. |
Danila Rozhevskii; Jie Zhu; Boyuan Zhao; | arxiv-cs.IR | 2022-05-06 |
439 | Is The Algorithmic Kadison-Singer Problem Hard? Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We study the following $\mathsf{KS}_2(c)$ problem: let $c \in\mathbb{R}^+$ be some constant, and $v_1,\ldots, v_m\in\mathbb{R}^d$ be vectors such that $\|v_i\|^2\leq \alpha$ for any $i\in[m]$ and $\sum_{i=1}^m \langle v_i, x\rangle^2 =1$ for any $x\in\mathbb{R}^d$ with $\|x\|=1$. |
Ben Jourdan; Peter Macgregor; He Sun; | arxiv-cs.CC | 2022-05-04 |
440 | Music Interpretation Analysis. A Multimodal Approach To Score-Informed Resynthesis of Piano Recordings Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: First, the main issue is identified in the understanding of how Music Information Processing (MIP) methods can take into consideration the influence of the acoustic context on the music performance. For this, a novel conceptual and mathematical framework named Music Interpretation Analysis (MIA) is presented. |
Federico Simonetta; | arxiv-cs.SD | 2022-05-02 |
441 | An Intelligent Music Genre Analysis Using Feature Extraction and Classification Using Deep Learning Techniques Related Papers Related Patents Related Grants Related Venues Related Experts View |
Hongdan Wang; Siti SalmiJamali; Chen Zhengping; Shan Qiaojuan; Ren Le; | Comput. Electr. Eng. | 2022-05-01 |
442 | Online EEG-Based Emotion Prediction and Music Generation for Inducing Affective States Related Papers Related Patents Related Grants Related Venues Related Experts View |
Kana Miyamoto; Hiroki Tanaka; Satoshi Nakamura; | IEICE Trans. Inf. Syst. | 2022-05-01 |
443 | College Music Education and Teaching Based on AI Techniques IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View |
Jing Wei; Marimuthu Karuppiah; A. Prathik; | Comput. Electr. Eng. | 2022-05-01 |
444 | Efficient Online Learning to Rank for Sequential Music Recommendation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Nevertheless, these approaches can suffer from slow convergence as a result of their random exploration component and get stuck in local minima as a result of their session-agnostic exploitation component. To overcome these limitations, we propose a novel online learning to rank approach which efficiently explores the space of candidate recommendation models by restricting itself to the orthogonal complement of the subspace of previous underperforming exploration directions. |
Pedro Dalla Vecchia Chaves; Bruno L. Pereira; Rodrygo L. T. Santos; | www | 2022-04-29 |
445 | Contrastive Learning with Positive-Negative Frame Mask for Music Representation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, those coarse-grained methods neglect some inessential or noisy elements at the frame level, which may be detrimental to the model to learn the effective representation of music. Towards this end, this paper proposes a novel Positive-nEgative frame mask for Music Representation based on the contrastive learning framework, abbreviated as PEMR. |
DONG YAO et. al. | www | 2022-04-29 |
446 | Music Enhancement Via Image Translation and Vocoding Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Consumer-grade music recordings such as those captured by mobile devices typically contain distortions in the form of background noise, reverb, and microphone-induced EQ. This paper presents a deep learning approach to enhance low-quality music recordings by combining (i) an image-to-image translation model for manipulating audio in its mel-spectrogram representation and (ii) a music vocoding model for mapping synthetically generated mel-spectrograms to perceptually realistic waveforms. |
Nikhil Kandpal; Oriol Nieto; Zeyu Jin; | arxiv-cs.SD | 2022-04-28 |
447 | Unaligned Supervision For Automatic Music Transcription in The Wild Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce NoteEM, a method for simultaneously training a transcriber and aligning the scores to their corresponding performances, in a fully-automated process. |
Ben Maman; Amit H. Bermano; | arxiv-cs.SD | 2022-04-28 |
448 | Large-Scale Multimodal Piano Music Identification Using Marketplace Fingerprinting Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: This paper studies the problem of identifying piano music in various modalities using a single, unified approach called marketplace fingerprinting. The key defining characteristic … |
Daniel Yang; Arya Goutam; Kevin Ji; T. Tsai; | Algorithms | 2022-04-26 |
449 | Music Emotion Recognition Based on Segment-level Two-stage Learning Related Papers Related Patents Related Grants Related Venues Related Experts View |
Na He; Sam Ferguson; | International Journal of Multimedia Information Retrieval | 2022-04-25 |
450 | SinTra: Learning An Inspiration Model from A Single Multi-track Music Segment Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose SinTra, an auto-regressive sequential generative model that can learn from a single multi-track music segment, to generate coherent, aesthetic, and variable polyphonic music of multi-instruments with an arbitrary length of bar. |
Qingwei Song; Qiwei Sun; Dongsheng Guo; Haiyong Zheng; | arxiv-cs.SD | 2022-04-21 |
451 | Uplifting Interviews in Social Science with Individual Data Visualization: The Case of Music Listening Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Designing a visual exploration tool allowing such an immersion is however difficult, because of the volume and heterogeneity of the listening data, the unequal visual literacy of the prospective users, or the interviewers’ potential lack of knowledge of the music listened to by the respondents. In this case study we discuss the design and evaluation of such a tool. |
ROBIN CURA et. al. | arxiv-cs.HC | 2022-04-14 |
452 | Deep Conditional Representation Learning for Drum Sample Retrieval By Vocalisation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this study, we investigated the potential of conditional autoencoder models to learn informative features for Drum Sample Retrieval by Vocalisation (DSRV). |
Alejandro Delgado; Charalampos Saitis; Emmanouil Benetos; Mark Sandler; | arxiv-cs.SD | 2022-04-10 |
453 | Genre-conditioned Acoustic Models for Automatic Lyrics Transcription of Polyphonic Music IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose to transcribe the lyrics of polyphonic music using a novel genre-conditioned network. |
Xiaoxue Gao; Chitralekha Gupta; Haizhou Li; | arxiv-cs.SD | 2022-04-07 |
454 | Late Multimodal Fusion for Image and Audio Music Transcription Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this respect, multimodal image and audio music transcription comprises the challenge of effectively combining the information conveyed by image and audio modalities. In this work, we explore this question at a late-fusion level: we study four combination approaches in order to merge, for the first time, the hypotheses regarding end-to-end OMR and AMT systems in a lattice-based search space. |
María Alfaro-Contreras; Jose J. Valero-Mas; José M. Iñesta; Jorge Calvo-Zaragoza; | arxiv-cs.MM | 2022-04-06 |
455 | A Novel Study on The MUSIC-type Imaging of Small Electromagnetic Inhomogeneities in The Limited-aperture Inverse Scattering Problem Related Papers Related Patents Related Grants Related Venues Related Experts View |
W. Park; | J. Comput. Phys. | 2022-04-01 |
456 | Machine Composition of Korean Music Via Topological Data Analysis and Artificial Neural Network Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present a way of machine composition that trains a machine the composition principle embedded in the given music data instead of directly feeding music pieces. |
Mai Lan Tran; Dongjin Lee; Jae-Hun Jung; | arxiv-cs.SD | 2022-03-29 |
457 | Iranian Modal Music (Dastgah) Detection Using Deep Neural Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we have implemented several deep neural networks to recognize Iranian modal music in seven highly correlated categories. |
Danial Ebrat; Farzad Didehvar; Milad Dadgar; | arxiv-cs.SD | 2022-03-29 |
458 | Subjective Evaluation of Deep Learning Models for Symbolic Music Composition Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a subjective method to evaluate AI-based music composition systems by asking questions related to basic music principles to different levels of users based on their musical experience and knowledge. |
Carlos Hernandez-Olivan; Jorge Abadias Puyuelo; Jose R. Beltran; | arxiv-cs.SD | 2022-03-28 |
459 | Automatic Song Translation for Tonal Languages Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose three criteria for effective AST — preserving meaning, singability and intelligibility — and design metrics for these criteria. |
FENFEI GUO et. al. | arxiv-cs.CL | 2022-03-24 |
460 | An Interactive Music Infilling Interface for Pop Music Composition Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This work builds a max patch for interactive music infilling application with different levels of control, including track density/polyphony/occupation rate and bar tonal tension control. |
Rui Guo; | arxiv-cs.SD | 2022-03-23 |
461 | CorpusVis: Visual Analysis of Digital Sheet Music Collections Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Bridging this gap, we contribute CorpusVis, an interactive visual workspace, enabling scalable and multi-faceted analysis. |
Matthias Miller; Julius Rauscher; Daniel A. Keim; Mennatallah El-Assady; | arxiv-cs.HC | 2022-03-23 |
462 | Music Students’ Perception Towards Music Distance Learning Education During COVID-19 Pandemic: Cross-Sectional Study in Jordan Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: During COVID-19 pandemic countries have faced various levels of COVID-19 infection rates, and millions of students are affected by changing the educational process. However, many … |
Nedal Nsairat; H. Fakhouri; R. Alsawalqa; Faten Hamad; | Int. J. Interact. Mob. Technol. | 2022-03-22 |
463 | Music Generation Using An LSTM Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Taking this into account, we provide a brief synopsis of the intuition, theory, and application of LSTMs in music generation, develop and present the network we found to best achieve this goal, identify and address issues and challenges faced, and include potential future improvements for our network. |
MICHAEL CONNER et. al. | arxiv-cs.SD | 2022-03-22 |
464 | TastePaths: Enabling Deeper Exploration and Understanding of Personal Preferences in Recommender Systems Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Recommender systems are ubiquitous and influence the information we consume daily by helping us navigate vast catalogs of information like music databases. However, their linear … |
S. PETRIDIS et. al. | 27th International Conference on Intelligent User Interfaces | 2022-03-22 |
465 | Music Trend Prediction Based on Improved LSTM and Random Forest Algorithm Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: As one of the entertainment consumption products, pop music attracts more and more people’s attention. In the context of big data, many pop music listeners can determine the … |
Xiangli Liu; | J. Sensors | 2022-03-22 |
466 | WeSinger: Data-augmented Singing Voice Synthesis with Auxiliary Losses Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we develop a new multi-singer Chinese neural singing voice synthesis (SVS) system named WeSinger. |
Zewang Zhang; Yibin Zheng; Xinhui Li; Li Lu; | arxiv-cs.SD | 2022-03-21 |
467 | Bi-Sampling Approach to Classify Music Mood Leveraging Raga-Rasa Association in Indian Classical Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper explores a novel framework to leverage the raga-rasa association in Indian classical Music to build an intelligent classifier and its application in music recommendation system based on user’s current mood and the mood they aspire to be in. |
MOHAN RAO B C et. al. | arxiv-cs.SD | 2022-03-13 |
468 | Multi-label Emotion Recognition from Indian Classical Music Using Gradient Descent SNN Model Related Papers Related Patents Related Grants Related Venues Related Experts View |
B. Tiple; Manasi S. Patwardhan; | Multimedia Tools and Applications | 2022-03-01 |
469 | Audience Experiences of A Volumetric Virtual Reality Music Video Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Music videos are short films that integrate songs and imagery and are produced for artistic and promotional purposes. Modern music videos apply various media capture techniques … |
Gareth W. Young; Néill O’dwyer; Matthew Moynihan; A. Smolic; | 2022 IEEE Conference on Virtual Reality and 3D User … | 2022-03-01 |
470 | Understanding Music Streaming Services Via Text Mining of Online Customer Reviews Related Papers Related Patents Related Grants Related Venues Related Experts View |
Jaemin Chung; Jiho Lee; Janghyeok Yoon; | Electron. Commer. Res. Appl. | 2022-03-01 |
471 | Jamming in MR: Towards Real-Time Music Collaboration in Mixed Reality Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Recent pandemic-related contact restrictions have made it difficult for musicians to meet in person to make music. As a result, there has been an increased demand for applications … |
R. SCHLAGOWSKI et. al. | 2022 IEEE Conference on Virtual Reality and 3D User … | 2022-03-01 |
472 | A Perceptual Measure for Evaluating The Resynthesis of Automatic Music Transcriptions Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose to distinguish the concept of performance from the one of interpretation, which expresses the artistic intention. |
Federico Simonetta; Federico Avanzini; Stavros Ntalampiras; | arxiv-cs.SD | 2022-02-24 |
473 | Flat Latent Manifolds for Human-machine Co-creation of Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: As a result, linearly interpolating in the latent space yields realistic and smooth musical changes that fit the type of machine–musician interactions we aim for. |
NUTAN CHEN et. al. | arxiv-cs.SD | 2022-02-23 |
474 | A Framework for Content-Based Search in Large Music Collections Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: We address the problem of scalable content-based search in large collections of music documents. Music content is highly complex and versatile and presents multiple facets that … |
Tiange Zhu; Raphaël Fournier-S’niehotta; P. Rigaux; Nicolas Travers; | Big Data Cogn. Comput. | 2022-02-23 |
475 | Enhancing Affective Representations of Music-Induced EEG Through Multimodal Supervision and Latent Domain Adaptation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper we extract efficient, personalized affective representations from EEG signals during music listening. |
Kleanthis Avramidis; Christos Garoufis; Athanasia Zlatintsi; Petros Maragos; | arxiv-cs.SD | 2022-02-20 |
476 | Interactive Multi-sensory and Volumetric Content Integration for Music Education Applications Related Papers Related Patents Related Grants Related Venues Related Experts View |
Chin-Ling Ho; Tsang-Gang Lin; Chan-Ru Chang; | Multimedia Tools and Applications | 2022-02-19 |
477 | Design and Innovation of Audio IoT Technology Using Music Teaching Intelligent Mode Related Papers Related Patents Related Grants Related Venues Related Experts View |
Lintao Li; Zhonglin Han; | Neural Computing and Applications | 2022-02-18 |
478 | Chord-Conditioned Melody Harmonization with Controllable Harmonicity Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: Melody harmonization has long been closely associated with chorales composed by Johann Sebastian Bach. Previous works rarely emphasised chorale generation conditioned on chord … |
Shangda Wu; Xiaobing Li; Maosong Sun; | arxiv-cs.SD | 2022-02-16 |
479 | Combined Angular Margin and Cosine Margin Softmax Loss for Music Classification Based on Spectrograms Related Papers Related Patents Related Grants Related Venues Related Experts View |
JINGXIAN LI et. al. | Neural Computing and Applications | 2022-02-12 |
480 | Learning Long-term Music Representations Via Hierarchical Contextual Constraints Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel approach to learn long-term symbolic music representations through contextual constraints. |
Shiqi Wei; Gus Xia; | arxiv-cs.SD | 2022-02-12 |
481 | Deep Performer: Score-to-Audio Music Performance Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Hence, we propose two new techniques for handling polyphonic inputs and providing a fine-grained conditioning in a transformer encoder-decoder model. |
Hao-Wen Dong; Cong Zhou; Taylor Berg-Kirkpatrick; Julian McAuley; | arxiv-cs.SD | 2022-02-12 |
482 | Audio Defect Detection in Music with Deep Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a convolutional network architecture following end-to-end encoder decoder configuration to develop detectors for two exemplary audio defects. |
Daniel Wolff; Rémi Mignot; Axel Roebel; | arxiv-cs.SD | 2022-02-11 |
483 | MusIAC: An Extensible Generative Framework for Music Infilling Applications with Multi-level Control Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a novel music generation framework for music infilling, with a user friendly interface. |
Rui Guo; Ivor Simpson; Chris Kiefer; Thor Magnusson; Dorien Herremans; | arxiv-cs.AI | 2022-02-11 |
484 | The HaMSE Ontology: Using Semantic Technologies to Support Music Representation Interoperability and Musicological Analysis Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this research, we propose HaMSE, an ontology capable of describing musical features that can assist musicological research. |
Andrea Poltronieri; Aldo Gangemi; | arxiv-cs.SD | 2022-02-11 |
485 | Barwise Compression Schemes for Audio-Based Music Structure Analysis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: More specifically, under the hypothesis that MSA is correlated with similarities occurring at the bar scale, this article introduces the use of linear and non-linear compression schemes on barwise audio signals. |
Axel Marmoret; Jérémy E. Cohen; Frédéric Bimbot; | arxiv-cs.SD | 2022-02-10 |
486 | An Intelligent Deep Learning-enabled Recommendation Algorithm for Teaching Music Students Related Papers Related Patents Related Grants Related Venues Related Experts View |
C. Tang; Jun Zhang; | Soft Computing | 2022-02-10 |
487 | MT3: Multi-Task Multitrack Music Transcription IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Unified framework for music transcription, jointly training a single model on six multi-instrument datasets and establishing a new SOTA for low-resource music transcription. |
Joshua P Gardner; Ian Simon; Ethan Manilow; Curtis Hawthorne; Jesse Engel; | iclr | 2022-02-08 |
488 | Research on The Architecture of Digital Song and Dance Costume Design System Based on Intelligent Deep Learning Algorithm Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: In order to improve the intelligence and reliability of song and dance costume design, this article combines the intelligent deep learning algorithm to improve the digital costume … |
Wenli Wang; | J. Sensors | 2022-02-07 |
489 | DanceFormer: Music Conditioned 3D Dance Generation with Parametric Motion Transformer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we reformulate it by a two-stage process, i.e., a key pose generation and then an in-between parametric motion curve prediction, where the key poses are easier to be synchronized with the music beats and the parametric curves can be efficiently regressed to render fluent rhythm-aligned movements. |
Buyu Li; Yongchi Zhao; Shi Zhelun; Lu Sheng; | aaai | 2022-02-07 |
490 | Musical Audio Similarity with Self-supervised Convolutional Neural Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We have built a music similarity search engine that lets video producers search by listenable music excerpts, as a complement to traditional full-text search. |
Carl Thomé; Sebastian Piwell; Oscar Utterbäck; | arxiv-cs.SD | 2022-02-04 |
491 | Melody Extraction from Polyphonic Music By Deep Learning Approaches: A Review Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, an attempt has been made to review the up-to-date data-driven deep learning approaches for melody extraction from polyphonic music. |
Gurunath Reddy M; K. Sreenivasa Rao; Partha Pratim Das; | arxiv-cs.SD | 2022-02-02 |
492 | ECF-MUSIC: An Empirical Characteristic Function Based Direction of Arrival (DOA) Estimation in The Presence of Impulsive Noise Related Papers Related Patents Related Grants Related Venues Related Experts View |
Mohsen Asghari; M. Zareinejad; S. M. Rezaei; H. Amindavar; | Digit. Signal Process. | 2022-02-01 |
493 | Partitura: A Python Package for Handling Symbolic Musical Data Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This demo paper introduces partitura, a Python package for handling symbolic musical information. |
Maarten Grachten; Carlos Cancino-Chacón; Thassilo Gadermaier; | arxiv-cs.SD | 2022-01-31 |
494 | Dual Learning Music Composition and Dance Choreography Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel extension, where we jointly model both tasks in a dual learning approach. |
Shuang Wu; Zhenguang Li; Shijian Lu; Li Cheng; | arxiv-cs.SD | 2022-01-28 |
495 | Analysis of Multimedia Teaching Path of Popular Music Based on Multiple Intelligence Teaching Mode Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Pop music multimedia is one of the popular digital pop music types. Based on the multiple intelligences teaching model, a multimedia multiple intelligences teaching method of … |
Dongfang Wang; | Adv. Multim. | 2022-01-28 |
496 | FIGARO: Generating Symbolic Music with Fine-Grained Artistic Control IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose the self-supervised description-to-sequence task, which allows for fine-grained controllable generation on a global level. |
Dimitri von Rütte; Luca Biggio; Yannic Kilcher; Thomas Hofmann; | arxiv-cs.SD | 2022-01-26 |
497 | Understanding and Compressing Music with Maximal Transformable Patterns Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a polynomial-time algorithm that discovers all maximal patterns in a point set, $D\subset\mathbb{R}^k$, that are related by transformations in a user-specified class, $F$, of bijections over $\mathbb{R}^k$. |
David Meredith; | arxiv-cs.LG | 2022-01-26 |
498 | Diversity in The Music Listening Experience: Insights from Focus Group Interviews Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Albeit several works have proposed new techniques for developing diversity-aware recommendations, little is known about how people perceive diversity while interacting with music recommendations. |
Lorenzo Porcaro; Emilia Gómez; Carlos Castillo; | arxiv-cs.HC | 2022-01-25 |
499 | Explainability in Music Recommender Systems IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this article, we discuss how explainability can be addressed in the context of MRSs. |
DARIUS AFCHAR et. al. | arxiv-cs.LG | 2022-01-25 |
500 | Multimedia Pop Music Teaching Model Integrating Semifinished Teaching Strategies Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: In order to improve the effect of popular music teaching, this paper combines the intelligent music frame feature recognition technology and the semifinished product teaching … |
Kangtan Dong; | Adv. Multim. | 2022-01-22 |
501 | A Survey of Music Emotion Recognition IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View |
Donghong Han; Yanru Kong; Jiayi Han; Guoren Wang; | Frontiers of Computer Science | 2022-01-22 |
502 | Kinit Classification in Ethiopian Chants, Azmaris and Modern Music: A New Dataset and CNN Benchmark Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we create EMIR, the first-ever Music Information Retrieval dataset for Ethiopian music. |
EPHREM A. RETTA et. al. | arxiv-cs.SD | 2022-01-20 |
503 | Can Machines Generate Personalized Music? A Hybrid Favorite-aware Method for User Preference Music Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: User preference music transfer (UPMT) is a new problem in music style transfer that can be applied to many scenarios but remains understudied. … |
Zhejing Hu; Yan Liu; Gong Chen; Yongxu Liu; | arxiv-cs.SD | 2022-01-20 |
504 | Opencpop: A High-Quality Open Source Chinese Popular Song Corpus for Singing Voice Synthesis IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper introduces Opencpop, a publicly available high-quality Mandarin singing corpus designed for singing voice synthesis (SVS). |
YU WANG et. al. | arxiv-cs.SD | 2022-01-19 |
505 | Unveiling Music Genre Structure Through Common-interest Communities Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Using a dataset of more than 90,000 metal music reviews written by over 9,000 users in a period of 15 years, we analyse the genre structure of metal music with the aid of review text information. |
Zhiheng Jiang; Hoai Nguyen Huynh; | arxiv-cs.SI | 2022-01-18 |
506 | A Novel Multi-Task Learning Method for Symbolic Music Emotion Recognition Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present a simple multi-task framework for SMER, which incorporates the emotion recognition task with other emotion-related auxiliary tasks derived from the intrinsic structure of the music. |
Jibao Qiu; C. L. Philip Chen; Tong Zhang; | arxiv-cs.SD | 2022-01-15 |
507 | Toward Piano Teaching Evaluation Based on Neural Network Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: With the rise of piano teaching in recent years, many people participated in the team of learning steel playing. However, expensive piano teaching fees and its unique one-to-one … |
Wanshu Luo; Bin Ning; | Sci. Program. | 2022-01-12 |
508 | Music2Video: Automatic Generation of Music Video with Fusion of Audio and Text Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Based on the fusion of text and audio, we create video whose content is consistent with the distinct modalities that are provided. |
Yoonjeon Kim; Joel Jang; Sumin Shin; | arxiv-cs.SD | 2022-01-11 |
509 | ‘Don’t Mess with My Algorithm’: Exploring The Relationship Between Listeners and Automated Curation and Recommendation on Music Streaming Services Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Given access to huge online collections of music on streaming platforms such as Spotify or Apple Music, users have become increasingly reliant on algorithmic recommender systems … |
Sophie Freeman; M. Gibbs; Bjørn Nansen; | First Monday | 2022-01-11 |
510 | Music Genre Classification Based on Auditory Image, Spectral and Acoustic Features Related Papers Related Patents Related Grants Related Venues Related Experts View |
Xin Cai; Hongjuan Zhang; | Multimedia Systems | 2022-01-10 |
511 | Machine Learning Model for Mapping of Music Mood and Human Emotion Based on Physiological Signals Related Papers Related Patents Related Grants Related Venues Related Experts View |
A. Garg; Vybhav Chaturvedi; A. Kaur; Vedansh Varshney; Anshu Parashar; | Multimedia Tools and Applications | 2022-01-08 |
512 | End-to-end Music Emotion Variation Detection Using Iteratively Reconstructed Deep Features IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View |
Richard Orjesek; R. Jarina; M. Chmulik; | Multimedia Tools and Applications | 2022-01-08 |
513 | Attentive Auto-encoder for Content-Aware Music Recommendation Related Papers Related Patents Related Grants Related Venues Related Experts View |
Le Li; Dan Tao; Chenwang Zheng; Ruipeng Gao; | CCF Transactions on Pervasive Computing and Interaction | 2022-01-03 |
514 | Measuring The Structural Complexity of Music: From Structural Segmentations to The Automatic Evaluation of Models for Music Generation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Composing musical ideas longer than motifs or figures is still rare in music generated by machine learning methods, a problem that is commonly referred to as the lack of long-term … |
Jacopo de Berardinis; A. Cangelosi; E. Coutinho; | IEEE/ACM Transactions on Audio, Speech, and Language … | 2022-01-01 |
515 | Difference-Frequency MUSIC for DOAs Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The direction-of-arrivals (DOAs) of plane waves in a high-frequency region are estimated without spatial aliasing using multi-frequency processing. The method exploits the … |
Yongsung Park; P. Gerstoft; Jeung-Hoon Lee; | IEEE Signal Processing Letters | 2022-01-01 |
516 | Augmented Tensor MUSIC for DOA Estimation Using Nested Acoustic Vector-Sensor Array Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Nested acoustic vector sensor (AVS) arrays have attracted growing interest, and their performance can be further improved by assembling AVSs with spatially separated (SS) … |
Xinghao Qu; Yi Lou; Yunjiang Zhao; Yinheng Lu; Gang Qiao; | IEEE Signal Processing Letters | 2022-01-01 |
517 | Personalized Synchronous Running Music Remix Procedure for Novice Runners Related Papers Related Patents Related Grants Related Venues Related Experts View |
NAN ZHUANG et. al. | International Conference on Evolutionary Computation | 2022-01-01 |
518 | Height Measurement with Meter Wave Polarimetric MIMO Radar: Signal Model and MUSIC-like Algorithm IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Inview of the advantages of polarization diversity and waveform diversity, polarimetric multiple-input multiple-output (MIMO) radar, which combines the two advantages, is a … |
Guimei Zheng; Yuwei Song; Chen Chen; | Signal Process. | 2022-01-01 |
519 | Computing Differential Galois Groups of Second-order Linear Q-difference Equations Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: We apply the differential Galois theory for difference equations developed by Hardouin and Singer to compute the differential Galois group for a second-order linear $q$-difference … |
Carlos E. Arreche; Yi Zhang; | Adv. Appl. Math. | 2022-01-01 |
520 | Merged Pitch Histograms and Pitch-duration Histograms Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: : The traditional pitch histogram and various features extracted from it play a pivotal role in music information retrieval. In the research on songs, especially applying pitch … |
Hui Liu; Tingting Xue; Tanja Schultz; | Signal Processing and Multimedia Applications | 2022-01-01 |
521 | Translating Melody to Chord: Structured and Flexible Harmonization of Melody with Transformer IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Recent deep learning approaches for melody harmonization have achieved remarkable performance by overcoming the uneven chord distributions of music data. However, most of these … |
Seungyeon Rhyu; Hyeonseok Choi; Sarah Kim; Kyogu Lee; | IEEE Access | 2022-01-01 |
522 | Multitrack Music Transformer: Learning Long-Term Dependencies in Music with Diverse Instruments Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Existing approaches for generating multitrack music with transformer models have been limited to either a small set of instruments or short music segments. This is partly due to … |
Hao-Wen Dong; K. Chen; S. Dubnov; Julian McAuley; Taylor Berg-Kirkpatrick; | ArXiv | 2022-01-01 |
523 | On The Use of Transformers for End-to-End Optical Music Recognition Related Papers Related Patents Related Grants Related Venues Related Experts View |
Antonio Ríos-Vila; J. Iñesta; Jorge Calvo-Zaragoza; | Iberian Conference on Pattern Recognition and Image Analysis | 2022-01-01 |
524 | Transfer Learning with Deep Neural Embeddings for Music Classification Tasks Related Papers Related Patents Related Grants Related Venues Related Experts View |
M. Modrzejewski; Piotr Szachewicz; P. Rokita; | International Conference on Artificial Intelligence and … | 2022-01-01 |
525 | An Overview of Automatic Piano Performance Assessment Within The Music Education Context Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: : Piano is one of the most popular instruments among music learners. Technologies to evaluate piano performances have been researched and developed in recent years rapidly, … |
Hyon Kim; Pedro Ramoneda; M. Miron; Xavier Serra; | International Conference on Computer Supported Education | 2022-01-01 |
526 | Self-Supervised Contrastive Learning for Singing Voices Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: This study introduces self-supervised contrastive learning to acquire feature representations of singing voices. To acquire robust representations in an unsupervised manner, … |
Hiromu Yakura; Kento Watanabe; Masataka Goto; | IEEE/ACM Transactions on Audio, Speech, and Language … | 2022-01-01 |
527 | Graph-Based Attentive Sequential Model With Metadata for Music Recommendation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Massive music data and diverse listening behaviors have caused great difficulties for existing methods in user-personalized recommendation scenarios. Most previous music … |
He Weng; Jianjiang Chen; Dongjing Wang; X. Zhang; Dongjin Yu; | IEEE Access | 2022-01-01 |
528 | BMNet-5: A Novel Approach of Neural Network to Classify The Genre of Bengali Music Based on Audio Features IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Music genre classification (MGC) is the process of putting genre labels on music by analyzing the sounds or words. With the rapid growth of music data repositories, MGC can be … |
KHAN MD HASIB et. al. | IEEE Access | 2022-01-01 |
529 | InverseMV: Composing Piano Scores with A Convolutional Video-Music Transformer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Therefore, we propose a novel attention-based model VMT (Video-Music Transformer) that automatically generates piano scores from video frames. |
Chin-Tung Lin; Mu Yang; | arxiv-cs.LG | 2021-12-31 |
530 | Evaluating Deep Music Generation Methods Using Data Augmentation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We analyse music samples generated by three models — SampleRNN, Jukebox, and DDSP — and employ a homogeneous framework across all methods to allow for objective comparison. |
TOBY GODWIN et. al. | arxiv-cs.SD | 2021-12-31 |
531 | Audio-to-symbolic Arrangement Via Cross-modal Music Representation Learning Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This is the audio-to-symbolic arrangement problem we tackle in this paper. |
Ziyu Wang; Dejing Xu; Gus Xia; Ying Shan; | arxiv-cs.SD | 2021-12-30 |
532 | On Skipping Behaviour Types in Music Streaming Sessions Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we investigate different behaviours during entire listening sessions with regards to the users’ session-based skipping activity. |
Francesco Meggetto; Crawford Revie; John Levine; Yashar Moshfeghi; | cikm | 2021-12-30 |
533 | Algorithmic Balancing of Familiarity, Similarity, & Discovery in Music Recommendations Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a multi-level ordered-weighted averaging based objective balancer to help maintain a healthy balance with respect to familiarity and discovery objectives, and conduct a series of offline evaluations and online AB tests, to demonstrate that despite the presence of strict trade-offs, we can achieve wins on both satisfaction and discover centric objectives. |
Rishabh Mehrotra; | cikm | 2021-12-30 |
534 | Exploring Key Competencies and Professional Development of Music Teachers in Primary Schools in The Era of Artificial Intelligence Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Artificial intelligence (AI) has introduced new era elements to the connotation of key competencies and professional development of music teachers in primary schools. The … |
Xiantong Zhao; Zhenjie Guo; Shanqin Liu; | Sci. Program. | 2021-12-23 |
535 | Artificial Intelligence Technology Assisted Music Teaching Design Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: With the continuous improvement of the global economic level and scientific level, information technology has penetrated into all fields of people’s life. Today, the strategy of … |
Dandan Dai; | Sci. Program. | 2021-12-21 |
536 | Generating Chords from Melody with Flexible Harmonic Rhythm and Controllable Harmonic Density Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose AutoHarmonizer, a harmonic density-controllable melody harmonization system with flexible harmonic rhythm. |
Shangda Wu; Yue Yang; Zhaowen Wang; Xiaobing Li; Maosong Sun; | arxiv-cs.SD | 2021-12-21 |
537 | User Experience Research on Social Robot Application Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: PurposeSocial robot, a subtype of robots that is designed for the various interactive services for human, which must deliver superior user experience (UX) by expressing human-like … |
Wei-Feng Tung; Jaileez Jara Santiago Campos; | Libr. Hi Tech | 2021-12-16 |
538 | EmotionBox: A Music-element-driven Emotional Music Generation System Using Recurrent Neural Network Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents EmotionBox -an music-element-driven emotional music generator that is capable of composing music given a specific emotion, where this model does not require a music dataset labeled with emotions. |
KAITONG ZHENG et. al. | arxiv-cs.SD | 2021-12-15 |
539 | Embedding-based Music Emotion Recognition Using Composite Loss Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose an embedding-based music emotion recognition approach that associates music samples with emotions in a common embedding space by considering both general emotional categories and fine-grained discrimination within each category. |
Naoki Takashima; Frédéric Li; Marcin Grzegorzek; Kimiaki Shirahama; | arxiv-cs.SD | 2021-12-14 |
540 | Vibrato Learning in Multi-Singer Singing Voice Synthesis Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Decent vibratos are a trait of good vocal training, often associated with perceived level of singing skill. In this paper we present a system for multi-singer singing voice … |
Ruolan Liu; Xue Wen; Chunhui Lu; Liming Song; June Sig Sung; | 2021 IEEE Automatic Speech Recognition and Understanding … | 2021-12-13 |
541 | Automatic Synthesis Technology of Music Teaching Melodies Based on Recurrent Neural Network Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Computer music creation boasts broad application prospects. It generally relies on artificial intelligence (AI) and machine learning (ML) to generate the music score that matches … |
Yingxue Zhang; Zhe Li; | Sci. Program. | 2021-12-09 |
542 | Learning Music Audio Representations Via Weak Language Supervision IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we pose the question of whether it may be possible to exploit weakly aligned text as the only supervisory signal to learn general-purpose music audio representations. |
Ilaria Manco; Emmanouil Benetos; Elio Quinton; Gyorgy Fazekas; | arxiv-cs.SD | 2021-12-08 |
543 | Context-Aware Music Recommender Systems for Groups: A Comparative Study Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Nowadays, recommender systems are present in multiple application domains, such as e-commerce, digital libraries, music streaming services, etc. In the music domain, these systems … |
Adrián Valera; Á. L. Murciego; M. M. García; | Inf. | 2021-12-07 |
544 | The Artificial Synesthete: Image-melody Translations with Variational Autoencoders Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Abstract This project presents a system of neural networks to translate between images and melodies. |
Karl Wienand; Wolfgang M. Heckl; | arxiv-cs.CV | 2021-12-06 |
545 | A Comprehensive Overview of AI-enabled Music Classification and Its Influence in Games Related Papers Related Patents Related Grants Related Venues Related Experts View |
Tiancheng Yang; S. Nazir; | Soft Computing | 2021-12-06 |
546 | Malakai: Music That Adapts to The Shape of Emotions Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: By leveraging these models and combining them with procedural algorithms from the last few decades, it is possible to create a dynamic song that composes music in real-time to accompany interactive experiences. |
ZACK HARRIS et. al. | arxiv-cs.MM | 2021-12-03 |
547 | Visual Storytelling of Song Ci and The Poets in The Social-cultural Context of Song Dynasty Related Papers Related Patents Related Grants Related Venues Related Experts View |
Wei Zhang; Qiang Ma; Rusheng Pan; Wei Chen; | Vis. Informatics | 2021-12-01 |
548 | Semi-supervised Music Emotion Recognition Using Noisy Student Training and Harmonic Pitch Class Profiles Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we intend to address the question: can we leverage semi-supervised learning techniques on music emotion recognition? |
Hao Hao Tan; | arxiv-cs.SD | 2021-12-01 |
549 | Score Transformer: Generating Musical Score from Note-level Representation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we explore the tokenized representation of musical scores using the Transformer model to automatically generate musical scores. |
Masahiro Suzuki; | arxiv-cs.SD | 2021-12-01 |
550 | A Knowledge-Based System for Children’s Music Teaching Strategies Based on The Inheritance of Local Music Culture in Southern Jiangsu Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: In order to analyze the feasibility of integrating Sunan music into children’s music teaching through research, a knowledge-based system for children’s music teaching strategies … |
Lina Su; Qing Jiang; | Int. J. Pattern Recognit. Artif. Intell. | 2021-11-30 |
551 | Expressive Communication: A Common Framework for Evaluating Developments in Generative Models and Steering Interfaces Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we investigate through a common framework how developments in both models and user interfaces are important for empowering co-creation where the goal is to create music that communicates particular imagery or ideas (e.g., as is common for other purposeful tasks in music creation like establishing mood or creating accompanying music for another media). |
Ryan Louie; Jesse Engel; Anna Huang; | arxiv-cs.HC | 2021-11-29 |
552 | Emotion Embedding Spaces for Matching Music to Stories Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, our goal is to help creators find music to match the emotion of their story. |
Minz Won; Justin Salamon; Nicholas J. Bryan; Gautham J. Mysore; Xavier Serra; | arxiv-cs.IR | 2021-11-26 |
553 | Semi-Supervised Music Tagging Transformer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present Music Tagging Transformer that is trained with a semi-supervised approach. |
Minz Won; Keunwoo Choi; Xavier Serra; | arxiv-cs.SD | 2021-11-26 |
554 | A-Muze-Net: Music Generation By Composing The Harmony Based on The Generated Melody Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a method for the generation of Midi files of piano music. |
Or Goren; Eliya Nachmani; Lior Wolf; | arxiv-cs.SD | 2021-11-25 |
555 | Towards Cross-Cultural Analysis Using Music Information Dynamics Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a framework that could be used to quantitatively compare music from different cultures by looking at these two aspects. |
Shlomo Dubnov; Kevin Huang; Cheng-i Wang; | arxiv-cs.SD | 2021-11-24 |
556 | Rhythm Is A Dancer: Music-Driven Motion Synthesis with Global Structure IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we present a music-driven motion synthesis framework that generates long-term sequences of human motions which are synchronized with the input beats, and jointly form a global structure that respects a specific dance genre. |
ANDREAS ARISTIDOU et. al. | arxiv-cs.GR | 2021-11-23 |
557 | Music Classification: Beyond Supervised Learning, Towards Real-world Applications Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this book, we present three lectures as follows: 1. |
Minz Won; Janne Spijkervet; Keunwoo Choi; | arxiv-cs.SD | 2021-11-22 |
558 | Comparing The Accuracy of Deep Neural Networks (DNN) and Convolutional Neural Network (CNN) in Music Genre Recognition (MGR): Experiments on Kurdish Music Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this research, we developed a dataset that contains 880 samples from eight different Kurdish music genres. |
Aza Zuhair; Hossein Hassani; | arxiv-cs.SD | 2021-11-22 |
559 | Music Students’ Use of Mobile Applications for Learning Purposes Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: This paper analyzes the use of mobile applications for learning purposes by music education department students. The survey design was used as the research method. The Mobile … |
Erkan Demirtaş; Sadik Özçelik; | International Journal of Modern Education Studies | 2021-11-21 |
560 | Catch-A-Waveform: Learning to Generate Audio from A Single Short Example IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Specifically, we present a GAN-based generative model that can be trained on one short audio signal from any domain (e.g. speech, music, etc.) and does not require pre-training or any other form of external supervision. |
Gal Greshler; Tamar Shaham; Tomer Michaeli; | nips | 2021-11-20 |
561 | How Does It Sound? Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we explore this problem and propose a novel system, called `RhythmicNet’, which takes as an input a video which includes human movements and generates a soundtrack for it. |
Kun Su; Xiulong Liu; Eli Shlizerman; | nips | 2021-11-20 |
562 | Video Background Music Generation with Controllable Music Transformer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we address the task of video background music generation. |
SHANGZHE DI et. al. | arxiv-cs.MM | 2021-11-16 |
563 | Symbolic Music Loop Generation with VQ-VAE Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, the objective is to generate a loop of 8 bars that can be used as a building block of music. |
Sangjun Han; Hyeongrae Ihm; Woohyung Lim; | arxiv-cs.SD | 2021-11-15 |
564 | Say What? Collaborative Pop Lyric Generation Using Multitask Transfer Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present a collaborative line-level lyric generation system that utilizes transfer-learning via the T5 transformer model, which, till date, has not been used to generate pop lyrics. |
Naveen Ram; Tanay Gummadi; Rahul Bhethanabotla; Richard J. Savery; Gil Weinberg; | arxiv-cs.CL | 2021-11-15 |
565 | A Transformer Generative Adversarial Network for Multi-track Music Generation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View |
CONG JIN et. al. | CAAI Trans. Intell. Technol. | 2021-11-09 |
566 | Theme Transformer: Symbolic Music Generation with Theme-Conditioned Transformer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We report on objective and subjective evaluations of variants of the proposed Theme Transformer and the conventional prompt-based baseline, showing that our best model can generate, to some extent, polyphonic pop piano music with repetition and plausible variations of a given condition. |
Yi-Jen Shih; Shih-Lun Wu; Frank Zalkow; Meinard Müller; Yi-Hsuan Yang; | arxiv-cs.SD | 2021-11-07 |
567 | Deep Embeddings and Section Fusion Improve Music Segmentation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Music segmentation algorithms identify the structure of a music recording by automatically dividing it into sections and determining which sections repeat and when. Since the … |
J. Salamon; Oriol Nieto; Nicholas J. Bryan; | 2021-11-07 | |
568 | Digital Audio Processing Tools for Music Corpus Studies Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The survey of extraction methods includes both a primer on signal processing and background theory on audio feature extraction. |
Johanna Devaney; | arxiv-cs.SD | 2021-11-06 |
569 | MT3: Multi-Task Multitrack Music Transcription IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, motivated by the promising results of sequence-to-sequence transfer learning for low-resource Natural Language Processing (NLP), we demonstrate that a general-purpose Transformer model can perform multi-task AMT, jointly transcribing arbitrary combinations of musical instruments across several transcription datasets. |
Josh Gardner; Ian Simon; Ethan Manilow; Curtis Hawthorne; Jesse Engel; | arxiv-cs.SD | 2021-11-04 |
570 | Learning To Generate Piano Music With Sustain Pedals Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: Recent years have witnessed a growing interest in research related to the detection of piano pedals from audio signals in the music information retrieval community. However, to … |
Joann Ching; Yi-Hsuan Yang; | arxiv-cs.SD | 2021-11-01 |
571 | ComposeInStyle: Music Composition with and Without Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View |
Sreetama Mukherjee; Manjunath Mulimani; | Expert Syst. Appl. | 2021-11-01 |
572 | A Novel 1D State Space for Efficient Music Rhythmic Analysis Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper proposes a new state space and a semi-Markov model for music time structure analysis. |
Mojtaba Heydari; Matthew McCallum; Andreas Ehmann; Zhiyao Duan; | arxiv-cs.SD | 2021-11-01 |
573 | Novel Datasets for Evaluating Song Popularity Prediction Tasks Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Estimating the success of a song before its release is an important music industry task. Current work uses audio descriptors to predict the success (popularity) of a song, where … |
M. Vötter; Maximilian Mayerl; Günther Specht; Eva Zangerle; | 2021 IEEE International Symposium on Multimedia (ISM) | 2021-11-01 |
574 | My Friends Also Prefer Diverse Music: Homophily and Link Prediction with User Preferences for Mainstream, Novelty, and Diversity in Music Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: The insights from this study will inform future work on social context-aware music recommendation, user modeling, and link prediction. |
Tomislav Duricic; Dominik Kowald; Markus Schedl; Elisabeth Lex; | arxiv-cs.SI | 2021-10-31 |
575 | Exploring Single-song Autoencoding Schemes for Audio-based Music Structure Analysis Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This work explores a piece-specific autoencoding scheme, in which a low-dimensional autoencoder is trained to learn a latent/compressed representation specific to a given song, which can then be used to infer the song structure. |
Axel Marmoret; Jérémy E. Cohen; Frédéric Bimbot; | arxiv-cs.SD | 2021-10-27 |
576 | Nonnegative Tucker Decomposition with Beta-divergence for Music Structure Analysis of Audio Signals Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This work proposes a multiplicative updates algorithm to compute NTD with the beta-divergence loss, often considered a better loss for audio processing. |
Axel Marmoret; Florian Voorwinden; Valentin Leplat; Jérémy E. Cohen; Frédéric Bimbot; | arxiv-cs.SD | 2021-10-27 |
577 | Unsupervised Source Separation By Steering Pretrained Music Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We use OpenAI’s Jukebox as the pretrained generative model, and we couple it with four kinds of pretrained music taggers (two architectures and two tagging datasets). |
Ethan Manilow; Patrick O’Reilly; Prem Seetharaman; Bryan Pardo; | arxiv-cs.SD | 2021-10-25 |
578 | Actions Speak Louder Than Listening: Evaluating Music Style Transfer Based on Editing Experience Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose an editing test to evaluate users’ editing experience of music generation models in a systematic way. |
Wei-Tsung Lu; Meng-Hsuan Wu; Yuh-Ming Chiu; Li Su; | arxiv-cs.SD | 2021-10-25 |
579 | Quantum Computer Music: Foundations and Initial Experiments Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: A bespoke generative music algorithm is presented, the Basak-Miranda algorithm, which leverages a property of quantum mechanics known as constructive and destructive interference to operate a musical Markov chain. |
Eduardo R. Miranda; Suchitra T. Bask; | arxiv-cs.ET | 2021-10-24 |
580 | Piano Online Teaching Based on Neural Network Technology Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: With the rapid growth of the number of digital music, how to find their favorite music in a large number of music has increasingly become a difficult problem for users. Based on … |
Yaokun Yang; | 2021 3rd International Conference on Artificial … | 2021-10-23 |
581 | SpecTNT: A Time-Frequency Transformer for Music Audio IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose SpecTNT, a Transformer-based architecture to model both spectral and temporal sequences of an input time-frequency representation. |
Wei-Tsung Lu; Ju-Chiang Wang; Minz Won; Keunwoo Choi; Xuchen Song; | arxiv-cs.SD | 2021-10-18 |
582 | MusicBERT: A Self-supervised Learning of Music Representation IF:3 Summary |