Results 271 to 280 of about 4,915,608 (331)
Some of the next articles are maybe not open access.
International Joint Conference on Artificial Intelligence
We introduce NotaGen, a symbolic music generation model aiming to explore the potential of producing high-quality classical sheet music. Inspired by the success of Large Language Models (LLMs), NotaGen adopts pre-training, fine-tuning, and reinforcement ...
Yashan Wang +9 more
semanticscholar +1 more source
We introduce NotaGen, a symbolic music generation model aiming to explore the potential of producing high-quality classical sheet music. Inspired by the success of Large Language Models (LLMs), NotaGen adopts pre-training, fine-tuning, and reinforcement ...
Yashan Wang +9 more
semanticscholar +1 more source
1991
Music suffers in discussion more than most arts. The difficulties of grasping the workings of an art whose materials of sound are intangible, elusive, and ephemeral are increased by the usual practice of employing physical and other alien metaphors to convey the activities of musical creation and appreciation.
openaire +1 more source
Music suffers in discussion more than most arts. The difficulties of grasping the workings of an art whose materials of sound are intangible, elusive, and ephemeral are increased by the usual practice of employing physical and other alien metaphors to convey the activities of musical creation and appreciation.
openaire +1 more source
Long-form music generation with latent diffusion
International Society for Music Information Retrieval ConferenceAudio-based generative models for music have seen great strides recently, but so far have not managed to produce full-length music tracks with coherent musical structure from text prompts.
Zach Evans +5 more
semanticscholar +1 more source
Joint Audio and Symbolic Conditioning for Temporally Controlled Text-to-Music Generation
International Society for Music Information Retrieval ConferenceWe present JASCO, a temporally controlled text-to-music generation model utilizing both symbolic and audio-based conditions. JASCO can generate high-quality music samples conditioned on global text descriptions along with fine-grained local controls ...
Or Tal +4 more
semanticscholar +1 more source
Generating Music from Flocking Dynamics
2012 American Control Conference (ACC), 2012We explore different approaches for generating music from the flocking dynamics of groups of mobile autonomous agents following a simple decentralized control rule. By developing software that links these dynamics to a set of sound wave generators, we study how each approach reflects sonically the transition to collective order and which produces ...
C. Huepe, R. F. Cadiz, M. Colasso
openaire +1 more source
Vision-to-Music Generation: A Survey
arXiv.orgVision-to-music Generation, including video-to-music and image-to-music tasks, is a significant branch of multimodal artificial intelligence demonstrating vast application prospects in fields such as film scoring, short video creation, and dance music ...
Zhaokai Wang +7 more
semanticscholar +1 more source
DITTO-2: Distilled Diffusion Inference-Time T-Optimization for Music Generation
International Society for Music Information Retrieval ConferenceControllable music generation methods are critical for human-centered AI-based music creation, but are currently limited by speed, quality, and control design trade-offs. Diffusion Inference-Time T-optimization (DITTO), in particular, offers state-of-the-
Zachary Novack +3 more
semanticscholar +1 more source
Seed-Music: A Unified Framework for High Quality and Controlled Music Generation
arXiv.orgWe introduce Seed-Music, a suite of music generation systems capable of producing high-quality music with fine-grained style control. Our unified framework leverages both auto-regressive language modeling and diffusion approaches to support two key music
Ye Bai +37 more
semanticscholar +1 more source
VMAs: Video-to-Music Generation via Semantic Alignment in Web Music Videos
IEEE Workshop/Winter Conference on Applications of Computer VisionWe present a framework for learning to generate background music from video inputs. Unlike existing works that rely on symbolic musical annotations, which are limited in quantity and diversity, our method leverages large-scale web videos accompanied by ...
Yan-Bo Lin +4 more
semanticscholar +1 more source
MusiConGen: Rhythm and Chord Control for Transformer-Based Text-to-Music Generation
International Society for Music Information Retrieval ConferenceExisting text-to-music models can produce high-quality audio with great diversity. However, textual prompts alone cannot precisely control temporal musical features such as chords and rhythm of the generated music. To address this challenge, we introduce
Y. Lan +3 more
semanticscholar +1 more source

