Production lessons from creating 12K+ AI tracks — prompt patterns that work, genres that stick, and the workflow behind a real music catalog.

You will know the prompt patterns, genre strategies, and production workflow behind 12,000+ AI-generated tracks.
TL;DR: After 12,000+ tracks on Suno, the lessons are not what most people expect. The best prompts are not the longest — they are the most specific about emotion and energy. Genre labels matter more than style descriptions. Vocal placement and instrument separation come from prompt structure, not luck. And the real value is not individual tracks — it is building a catalog system with metadata, hosting, and distribution. Here is what actually works.
12,000 tracks sounds like a lot. It felt like a lot the first time I looked at the number. But the more useful figure is this: I have 65 tracks indexed in data/inventories/frankx/music.json, 61 of them self-hosted on Vercel Blob, with zero unnamed tracks and zero tracks without genre tags.
That gap between 12,000 generated and 65 cataloged tells you everything about what AI music production actually is. It is not a creative flow where every output is a gem. It is a generation-and-curation process, and the curation infrastructure matters at least as much as the generation quality.
Most people who use Suno build up a chaotic library of tracks with names like "Energetic Pop 3" and no systematic way to find, license, or distribute them. I built a catalog system. That is the real lesson.
After running scripts/enrich-suno-tracks.mjs against my full catalog, the genre distribution in my indexed tracks breaks down like this:
| Genre | Tracks |
|---|---|
| Pop | 14 |
| Vocal | 14 |
| Dance | 12 |
| Neoclassical | 12 |
| Electronic | 12 |
| Orchestral | 9 |
These are not the genres I set out to produce. They are the genres where Suno's output quality consistently cleared the bar for catalog inclusion. Pop and Vocal made the cut most often because the production quality is most consistent there — Suno handles clear lead vocals over straightforward arrangements better than most other configurations.
Neoclassical was a surprise. I started prompting in that space because I wanted music for ambient listening, and the success rate was high enough that it became a real focus. 12 cataloged Neoclassical tracks represent maybe 200 generation attempts — a roughly 6% catalog-inclusion rate, which is actually above average.
Dance and Electronic have the worst ratio of generations to catalog inclusions — lots of attempts, fewer that hold up. The model handles rhythm and energy well but struggles with transitions and builds in longer electronic formats.
After generating this volume, the patterns are clear. Here is what I have empirically established:
The single most impactful change I made was moving genre to the first position in every prompt. Not as a style description — as a precise genre label.
Wrong: "Create an upbeat electronic track with pop sensibility" Right: "Dance Pop"
The model responds differently to genre labels versus style descriptions. Genre labels activate trained patterns. Style descriptions trigger interpretation. Interpretation introduces variance. Variance means inconsistency.
I run every serious prompt through a five-layer structure:
Layer 1 — Genre Tag: Precise label, no descriptions. "Neoclassical", "Dance Pop", "Orchestral Cinematic".
Layer 2 — Energy Modifier: Single word. "Driving", "Ethereal", "Melancholic", "Triumphant". This sets the emotional center of mass.
Layer 3 — Instrument Specificity: Name the instruments that matter. Not "piano and strings" — "solo piano, string quartet, cello lead". Specificity produces separation.
Layer 4 — Structural Directive: Where the energy goes. "Builds to a climax at the two-minute mark", "Stays sparse throughout", "Drops to half-time in the bridge".
Layer 5 — Production Detail: One or two specific production characteristics. "Dry room, close-mic feel", "Wide reverb on the leads", "Minimal compression on the low end".
A full prompt using this architecture: Neoclassical. Melancholic. Solo piano, string quartet, cello lead. Sparse opening builds to full ensemble at 90 seconds. Close-mic piano, wide reverb on strings.
That prompt produces usable output at roughly 35-40%. Without the structure, the same creative intent might produce usable output 5-10% of the time.
Vocal placement is the most requested improvement I hear from other Suno users. Where does the lead vocal sit in the mix? When does it enter? How present is it?
My findings after this volume: vocal placement responds most to song section markers in lyrics mode. If you structure your lyrics with [Verse 1], [Pre-Chorus], [Chorus], [Bridge] markers, the model uses those as structural anchors and places the vocal accordingly.
Without markers, the model improvises structure. Sometimes that works. At catalog scale, it does not work consistently enough.
The other lever is the presence modifier in the prompt. "Prominent vocals", "Whispered lead", "Choir-backed lead", "Solo vocal with minimal backing" — these work. "Beautiful singing" does not work. The model needs production language, not aesthetic language.
My full workflow from a generation session to a cataloged track:
Step 1 — Generation Run prompts in sets of four. Suno generates two variants per prompt. I evaluate eight outputs per creative direction and select the best one or two.
Step 2 — Metadata Enrichment
scripts/enrich-suno-tracks.mjs pulls title, genre, duration, and cover art from each track's Suno page. The sunoId becomes the primary key: {sunoId}.mp3.
Step 3 — Vercel Blob Upload
scripts/upload-products-to-blob.mjs (adapted for music) pushes each MP3 to Vercel Blob at path music/{sunoId}/{sunoId}.mp3. 61 tracks currently live at this path. Suno CDN pattern (https://cdn1.suno.ai/{sunoId}.mp3) provides a backup source.
Step 4 — Catalog Entry
Each track gets a JSON entry in data/inventories/frankx/music.json with: sunoId, title, genre array, mood tags, duration, blob URL, cover art URL, album assignment (if any), and license status.
Step 5 — n8n Sync
The "Music Catalog Sync" workflow (ID: AEalmmG7xGSmq6Wh) runs daily at 8am UTC and on-demand via webhook at /music-sync. It checks for new Suno releases on my profile (@frankx — not @frankxmusic, which is empty) and queues enrichment automatically.
Step 6 — Player Integration
The custom audio player at lib/player-context.tsx reads from the catalog JSON. components/music/TrackCard.tsx renders each entry. The global player persists across navigation. You can listen to the catalog at /music.
Generating without a selection framework. If you do not know what "good enough for catalog" means before you generate, you will either include too much (degrading catalog quality) or too little (burning out on curation). I target: would I pay $1 to license this for a YouTube video? Yes = catalog. No = discard.
Style over structure in prompts. "Cinematic and emotional" generates nothing useful. "Orchestral, triumphant, full brass section, builds over 3 minutes with a climax on the downbeat of measure 48" generates something the model can actually attempt.
Not capturing the sunoId immediately. Suno's URLs contain the track ID. If you close a session without recording it, the track is findable but requires extra steps. I built a browser bookmark that extracts and copies the sunoId from any Suno URL as the first step of any evaluation session.
Treating every generation as a single. The economics of AI music only make sense at catalog scale. A single track has minimal commercial value. A catalog of 60+ tracks in consistent genres with proper metadata has licensing potential, sync potential, streaming royalty potential. Catalog > singles, always.
Ignoring album architecture. I have five albums in draft state in data/albums.json: Golden Frequencies, Arcanea Soundtrack, Electric Soul, Vibe State, and Tech House Sessions. Album framing changes how catalog tracks are perceived. The same track sounds different when it is "Track 7 on Vibe State" versus "a random MP3 in my downloads folder."
The 5-Layer Prompt Architecture I use for Suno is documented in expanded form in the Prompt Library. The music production entries include genre-specific templates, energy modifier combinations that work for each genre, and the production detail vocabulary that produces consistent results.
The research on AI creative tools includes comparative analysis of Suno's output quality across genre families based on my generation history. The neoclassical and orchestral findings there are based on the same catalog data referenced in this article.
For the technical side of the Suno prompt engineering approach, the Suno prompt engineering complete guide goes deeper on the structural mechanics — why genre labels work the way they do, how the model interprets production vocabulary, and advanced techniques for specific subgenres.
Streaming royalties from a single track are negligible. Sync licensing from a single track requires luck and connections. A catalog changes the math.
With 61 self-hosted tracks across six genre families:
I built five draft albums because album releases give existing tracks new context without requiring new generation. The same 12 neoclassical tracks that are currently scattered across data/inventories/frankx/music.json become a cohesive release when packaged under a single cover art, sequenced deliberately, and distributed as a body of work.
The production leverage is significant: generation cost is already paid. The incremental cost of album packaging, metadata, and distribution is hours of work, not days. The output is a professional release.
The number is not the point. What matters is what happens to your intuition when you generate at that volume:
You stop being surprised by individual outputs and start reading patterns. You know within the first four seconds whether a track is on the right trajectory. You know which prompt changes will shift the output in which direction. You stop fighting the model's tendencies and start working with them.
Neoclassical at this volume means I know exactly which instrument combinations produce the warmest string tone. Dance Pop means I know which energy modifiers produce the specific driving rhythm that sits right in a mix. This is not talent. It is calibration from volume.
The other thing 12,000 tracks teaches: generation is cheap, curation is expensive, and infrastructure is the only thing that makes scale sustainable. The workflow described in this article exists because I got buried under uncatalogued tracks at around 500 generations and had to build a system or lose everything I had made.
Build the infrastructure at 50 tracks. Not at 5,000.
Move genre to the first position and use precise genre labels instead of style descriptions. "Dance Pop" will outperform "upbeat electronic with pop sensibility" on consistency every time. The model was trained on genre labels — use the vocabulary it knows.
I use a single filter question: would I pay $1 to license this for a video? If yes, it goes into the catalog. If not, it is discarded. This creates a consistent quality bar that does not drift with mood or fatigue. Apply it within the first 30 seconds of listening — that is enough to judge production quality and arrangement integrity.
Yes, for anyone building a real catalog. Suno CDN links are not guaranteed permanent. Self-hosting gives you control over availability, streaming performance, and player integration. The setup is a one-time build — the scripts/upload-products-to-blob.mjs script handles uploads, and the custom player at lib/player-context.tsx handles playback. After initial setup, adding new tracks is a single command.
Genre labels in Suno prompts trigger trained associations — the model has processed thousands of labeled examples and learned to reproduce the production characteristics associated with each label. Style descriptions require the model to interpret what you mean, which introduces variance. Genre labels are closer to direct instruction; style descriptions are closer to suggestion.
The music catalog system is a direct application of ACOS infrastructure principles: persistent data storage (data/inventories/frankx/music.json), automated enrichment and sync (n8n workflow), and systematic output management (Vercel Blob hosting + custom player). The same architecture that makes AI work sessions coherent makes a music catalog coherent. Treat every creative output as data, not just art.
Step-by-step guide to setting up ACOS, creating your first agent, and shipping real products with AI.
Start buildingDownload AI architecture templates, multi-agent blueprints, and prompt engineering patterns.
Browse templatesConnect with creators and architects shipping AI products. Weekly office hours, shared resources, direct access.
Join the circleRead on FrankX.AI — AI Architecture, Music & Creator Intelligence
Weekly field notes on AI systems, production patterns, and builder strategy.

Transform your ideas into cinematic soundscapes, meditation music, and motivation anthems using AI. Learn the 5-Layer Prompt Architecture and frequency science for transformative audio.
Read article
Move beyond "hit and hope" generation. A systematic workflow for using Suno AI to produce radio-ready tracks, from emotion mapping to stem mixing.
Read article
An integrated platform combining AI music production, multi-LLM orchestration, and intelligent agent coordination. Built on Claude, GPT, Gemini, Grok, and Llama.
Read article