Welcome to ACE-Step Discussions! #89
Replies: 4 comments 1 reply
-
|
Hello ACE-Step Team, thank you all for creating such an awesome foundation model for the community to try out and learn from. I really appreciate all of your hard work. I have used the free version of Suno a lot, so seeing ACE-Step was exciting to see and try out. Feedback from user experience:
THANK YOU AGAIN FOR ALL THAT YOU DO! |
Beta Was this translation helpful? Give feedback.
-
|
Just saying: I love ACE-Step! Great work! The ComfyUI workflow works great for my potato hardware and I would love to see a easy way to train LORA in ComfyUI. So team, keep on rockin'! You're doing fantastic work! |
Beta Was this translation helpful? Give feedback.
-
|
Hello, I’m a music novice. Thank you very much to you and your team for sharing the ACE-Step model—a foundational tool dedicated to music generation and song synthesis—which provides a creative platform for people who aspire to create music. I’d like to ask how to precisely control specific lyrics or time segments. For example, having a gentle, classical-style piano solo from 15s to 26s in a song... Or emphasizing a passionate and uplifting mood during the performance of certain lyrics... |
Beta Was this translation helpful? Give feedback.
-
|
Greetings! Apart from my GitHub activities as an open source software maintainer, I've been a singer, songwriter, and musician for a while and used a lot of tools along the way, from pen & paper to POKEY chips and Amiga Trackers, to Cubase and GarageBand, to Absynth and SunVox. I even published some music apps ("FretPet" and "ChordCalc") to deconstruct and play with music in realtime. I've always enjoyed exploring the available tools of the age, especially now in the Space Age. As a prolific creator of melodies and lyrics I've been looking for tools to develop more complete compositions with accompaniment, without having to perform all the parts myself, without having to coordinate musician schedules, and within a short time. One wants to realize ideas while they are still fresh, even when resources are constrained. In general my experience with ACE Step has been one of continuous surprise and amazement. There is always something a bit off about the result, but it gets closer than what seems should be possible for such a small model. I'm excited to have access to a model like this that can realize such a diverse range of musical styles. If only it could run faster with MLX in macOS…. What can ACE Step do?The ACE Step v1 3.5B model does so much at once, creating a complete piece, apparently using a diffusion-like process. That makes it really mind-blowing as a tool to turn a set of lyrics into a styled piece of music, to generate ideas as you tweak the lyrics and styles. Because it generates everything as a mix it doesn't allow the user to make very specific adjustments, so for the composer who wants to take further steps this model provides a great tool to create scratch tracks and endless variants. Given its very specific bag of tricks it makes me wonder how this and other models can be applied when wrapped up as plugins for apps like Logic, GarageBand, and others, or if one could just vibe-code a multi-track timeline for laying down generated sections, creating loops, with cut-and-paste, effects, EQ, etc. Is that too much to ask from Gradio? More audio tricksIt's not clear, but it seems that ACE Step v1 cannot be prompted to do things like isolate the drum track from a song input, or create a new, isolated bass guitar track to accompany a given piece of music. However, models very much like ACE Step v1 should be able to do all these tricks, and more. I believe that NVIDIA has some large models that do this kind of thing. It gives hope for other capabilities to filter down to our common and open source toolsets. ACE Step as CodecAnother very interesting thing about such models is that a 4 minute composition can now be effectively encoded into a tiny 1K JSON file. It takes some time and energy for the model to "decode" that JSON into an audio track, but its deterministic result makes it effectively a compression algorithm. Probably there's no chance to run it backwards, to turn an unfamiliar audio track into a set of concise prompts. But… what if it could? Further Model Development
UI Development
I would enjoy hearing from others who have ideas about how to fit this model into a common production workflow, and how we might get models like this incorporated into plugins for our favorite music apps. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
👋 Welcome to Our ACE-Step Community Hub!
We're using Discussions as a space to explore everything about this model - from trying demos to advanced training! Here you can:
Popular Topics:
🎮 Demo & Applications
⚙️ Training & Tuning
🧩 Advanced Techniques
Getting Started:
Remember: Every question helps someone else learn too! ✨
Beta Was this translation helpful? Give feedback.
All reactions