Results 271 to 280 of about 4,915,608 (331)
Some of the next articles are maybe not open access.

NotaGen: Advancing Musicality in Symbolic Music Generation with Large Language Model Training Paradigms

International Joint Conference on Artificial Intelligence
We introduce NotaGen, a symbolic music generation model aiming to explore the potential of producing high-quality classical sheet music. Inspired by the success of Large Language Models (LLMs), NotaGen adopts pre-training, fine-tuning, and reinforcement ...
Yashan Wang   +9 more
semanticscholar   +1 more source

Musical Generation

1991
Music suffers in discussion more than most arts. The difficulties of grasping the workings of an art whose materials of sound are intangible, elusive, and ephemeral are increased by the usual practice of employing physical and other alien metaphors to convey the activities of musical creation and appreciation.
openaire   +1 more source

Long-form music generation with latent diffusion

International Society for Music Information Retrieval Conference
Audio-based generative models for music have seen great strides recently, but so far have not managed to produce full-length music tracks with coherent musical structure from text prompts.
Zach Evans   +5 more
semanticscholar   +1 more source

Joint Audio and Symbolic Conditioning for Temporally Controlled Text-to-Music Generation

International Society for Music Information Retrieval Conference
We present JASCO, a temporally controlled text-to-music generation model utilizing both symbolic and audio-based conditions. JASCO can generate high-quality music samples conditioned on global text descriptions along with fine-grained local controls ...
Or Tal   +4 more
semanticscholar   +1 more source

Generating Music from Flocking Dynamics

2012 American Control Conference (ACC), 2012
We explore different approaches for generating music from the flocking dynamics of groups of mobile autonomous agents following a simple decentralized control rule. By developing software that links these dynamics to a set of sound wave generators, we study how each approach reflects sonically the transition to collective order and which produces ...
C. Huepe, R. F. Cadiz, M. Colasso
openaire   +1 more source

Vision-to-Music Generation: A Survey

arXiv.org
Vision-to-music Generation, including video-to-music and image-to-music tasks, is a significant branch of multimodal artificial intelligence demonstrating vast application prospects in fields such as film scoring, short video creation, and dance music ...
Zhaokai Wang   +7 more
semanticscholar   +1 more source

DITTO-2: Distilled Diffusion Inference-Time T-Optimization for Music Generation

International Society for Music Information Retrieval Conference
Controllable music generation methods are critical for human-centered AI-based music creation, but are currently limited by speed, quality, and control design trade-offs. Diffusion Inference-Time T-optimization (DITTO), in particular, offers state-of-the-
Zachary Novack   +3 more
semanticscholar   +1 more source

Seed-Music: A Unified Framework for High Quality and Controlled Music Generation

arXiv.org
We introduce Seed-Music, a suite of music generation systems capable of producing high-quality music with fine-grained style control. Our unified framework leverages both auto-regressive language modeling and diffusion approaches to support two key music
Ye Bai   +37 more
semanticscholar   +1 more source

VMAs: Video-to-Music Generation via Semantic Alignment in Web Music Videos

IEEE Workshop/Winter Conference on Applications of Computer Vision
We present a framework for learning to generate background music from video inputs. Unlike existing works that rely on symbolic musical annotations, which are limited in quantity and diversity, our method leverages large-scale web videos accompanied by ...
Yan-Bo Lin   +4 more
semanticscholar   +1 more source

MusiConGen: Rhythm and Chord Control for Transformer-Based Text-to-Music Generation

International Society for Music Information Retrieval Conference
Existing text-to-music models can produce high-quality audio with great diversity. However, textual prompts alone cannot precisely control temporal musical features such as chords and rhythm of the generated music. To address this challenge, we introduce
Y. Lan   +3 more
semanticscholar   +1 more source

Home - About - Disclaimer - Privacy