Eachlabs | AI Workflows for app builders
mureka-generate-song

MUREKA

Mureka Generate Song is a full music generation model that creates complete songs from prompts, lyrics, or structured inputs.

Avg Run Time: 65.000s

Model Slug: mureka-generate-song

Playground

Input

Advanced Controls

Output

Example Result

Preview and download your result.

Calculated using formula: 2 * 0.05. Cost per execution: $0.1000

API & SDK

Create a Prediction

Send a POST request to create a new prediction. This will return a prediction ID that you'll use to check the result. The request should include your model inputs and API key.

Get Prediction Result

Poll the prediction endpoint with the prediction ID until the result is ready. The API uses long-polling, so you'll need to repeatedly check until you receive a success status.

Readme

Table of Contents
Overview
Technical Specifications
Key Considerations
Tips & Tricks
Capabilities
What Can I Use It For?
Things to Be Aware Of
Limitations

Overview

mureka-generate-song — Text-to-Audio AI Model

Mureka Generate Song, part of Mureka's advanced mureka family from Kunlun Tech, revolutionizes AI music generation by creating complete, professional songs from text prompts, lyrics, or reference tracks, solving the challenge of rapid, high-quality music production without traditional studio resources. This mureka-generate-song model stands out with its MusiCoT technology, which employs chain-of-thought reasoning to plan song structures like verses, choruses, and bridges for superior musical coherence. Developers and creators searching for "AI song generator with stems" or "text to music AI model" find mureka-generate-song ideal for generating full tracks up to 240 seconds with vocals, instrumentation, and exportable stems.

Technical Specifications

What Sets mureka-generate-song Apart

The mureka-generate-song model differentiates itself in the crowded AI music space through specific capabilities like MusiCoT reasoning for structured compositions and multi-stem exports, enabling professional-grade outputs that integrate seamlessly into DAWs. It supports multi-modal inputs including text prompts, hummed melodies, reference tracks, and lyrics, with generation times typically under a minute for tracks up to 240 seconds in formats like MP3, WAV, and FLAC.

  • MusiCoT Chain-of-Thought Planning: The model reasons step-by-step on time signatures, BPM, and sections like verses and choruses before audio synthesis. This delivers coherent, logically structured songs that feel professionally composed, unlike basic text-to-audio tools.
  • Multi-Track Stem Separation: Exports individual stems for vocals, drums, bass, and synths compatible with Ableton or Logic Pro. Users gain granular editing control, perfect for "AI music generator with stems" workflows in professional production.
  • Advanced Vocal Features with 10+ Languages: Includes voice cloning, multi-voice duets, and synthesis in languages like English, Chinese, Japanese, and Spanish. This enables authentic global tracks, setting it apart for creators needing "AI song generator multilingual vocals."
  • Multi-Modal Inputs and Editing: Accepts hummed melodies, YouTube links, or reference audio for style matching, plus regional editing and extensions. This flexibility supports iterative creation for "custom AI music from reference tracks."

Key Considerations

  • Use detailed style prompts combining genre, subgenre, mood, tempo/energy, instrumentation, structure, and mix notes for best adherence to vision
  • Provide lyrics with structural tags like [verse] or [chorus] to improve song organization and coherence
  • Balance quality settings to access advanced features and achieve professional results
  • Start with basic modes for quick ideation, then iterate to custom prompts for refinement
  • Avoid overly vague prompts like single genres; specificity reduces inconsistencies in vocals and arrangement
  • Trade-off: Higher detail in prompts yields better quality but may increase generation time
  • Prompt engineering tip: Structure as "Genre + mood + tempo + instruments + structure" e.g., "indie pop, bittersweet hopeful, mid-tempo, warm synths + guitar, verse-chorus with big hook"

Tips & Tricks

How to Use mureka-generate-song on Eachlabs

Access mureka-generate-song seamlessly on Eachlabs via the Playground for instant testing with text prompts, lyrics, hummed melodies, or reference audio; integrate through the robust API or SDK for apps needing "mureka-generate-song API" functionality. Specify parameters like model version (e.g., O2), vocals, language, and duration up to 240 seconds to produce high-fidelity MP3/WAV outputs with stems, typically in under a minute.

---

Capabilities

  • Generates complete songs with melody, harmony, expressive natural vocals, and professional instrumentation from lyrics + prompts
  • Handles long-form music up to 6 minutes with structural coherence and emotional diversity
  • Precise control via text: genre, mood, tempo, vocal character, energy, atmosphere
  • Special features: Fine-grained style control per section (intro/verse/chorus), short-music mode for clips, reference audio conditioning
  • High-quality outputs: Smoother harmonies, balanced mixes, emotionally rich vocals superior to prior versions like V7.5
  • Versatile across genres, moods; adaptable to instrumental-only or sung tracks

What Can I Use It For?

Use Cases for mureka-generate-song

Music producers building personalized soundtracks use mureka-generate-song's custom voice cloning to train models on their catalog of 100-200 tracks, generating new songs that match their signature style with stems for DAW remixing—ideal for artists seeking "AI music generator with voice cloning."

Marketers creating ad jingles input lyrics and a reference track like "upbeat pop with synths and female vocals in Spanish," receiving a full 4-minute song with separated stems in seconds, streamlining campaigns without hiring composers.

Developers integrating "mureka-generate-song API" into apps hum a melody motif via mobile input, then extend it into a complete electronic track with multi-language vocals, enabling scalable music features for gaming or social platforms.

Content creators for YouTube or podcasts generate ambient scores by prompting "dark cinematic ambience with piano and strings," editing specific sections and exporting MIDI for further customization, perfect for "text to music AI model" needs in video production.

Things to Be Aware Of

  • Experimental behaviors: Vocals and arrangements improve with V8 over V7.5, but may require prompt tweaks for perfect balance
  • Known quirks: Outputs adhere closely to prompts but can vary in energy/tempo without specifics
  • Performance considerations: Efficient for API integration; longer tracks maintain coherence via advanced tokenization
  • Resource requirements: Runs via API endpoints; balance settings needed for full features
  • Consistency factors: Multiple generations recommended for best takes, as iteration tightens results
  • Positive user feedback themes: Praised for natural vocals, emotional depth, and ease from lyrics to full song
  • Common concerns: Less ideal for purely instrumental loops without lyrics; may need refinement for niche subgenres

Limitations

  • Lacks publicly detailed parameter counts or exact training data, limiting deep technical replication
  • Primarily excels with lyrics inputs; instrumental-only or description-based generation may be less structured without them
  • Potential inconsistencies in highly experimental styles, requiring multiple iterations for optimal results