DREAMOMNI2
DreamOmni2/Edit is a multimodal model for precise and creative image editing guided by text and visual inputs.
Avg Run Time: 40.000s
Model Slug: dreamomni2-edit
Playground
Input
Output
Example Result
Preview and download your result.
API & SDK
Create a Prediction
Send a POST request to create a new prediction. This will return a prediction ID that you'll use to check the result. The request should include your model inputs and API key.
Get Prediction Result
Poll the prediction endpoint with the prediction ID until the result is ready. The API uses long-polling, so you'll need to repeatedly check until you receive a success status.
Readme
Overview
dreamomni2-edit — Image Editing AI Model
Developed by Bytedance as part of the DreamOmni2 family, dreamomni2-edit is a multimodal image editing model that combines text prompts with visual references to deliver precise, context-aware edits. Unlike traditional image-to-image AI models that rely on text alone, dreamomni2-edit accepts both textual instructions and reference images simultaneously, enabling creators to specify exactly what should change and how the result should look in a single operation. This dual-input approach eliminates the guesswork and iteration cycles common in AI image editing workflows, making it ideal for professionals who need reliable, repeatable results.
The model excels at understanding semantic relationships within images—it knows the difference between a product, its background, and lighting conditions, allowing edits that preserve object integrity while transforming surroundings. Whether you're adjusting product photography for e-commerce, refining creative compositions, or exploring design variations, dreamomni2-edit delivers edits that respect the original image's structure while applying your creative vision.
Technical Specifications
What Sets dreamomni2-edit Apart
dreamomni2-edit distinguishes itself through capabilities specifically designed for professional image editing workflows:
- Dual-input guidance: Accepts text prompts and visual references in parallel, allowing you to specify both the editing instruction ("change the background to a sunset") and the visual style ("match this reference image") simultaneously. This reduces the trial-and-error cycles typical of text-only image-to-image AI models.
- Semantic-aware transformations: Preserves object identity, spatial relationships, and product integrity during edits—critical for e-commerce photography where maintaining accurate product representation while changing backgrounds, lighting, or settings is non-negotiable.
- Inpainting and outpainting: Supports both localized region editing (inpainting) and creative image expansion (outpainting) within the same model, enabling workflows from precise touch-ups to full compositional reimagining.
- High-resolution output: Generates edits up to 4K resolution with support for multiple aspect ratios, ensuring results are suitable for print, e-commerce platforms, and high-fidelity creative work.
Processing is optimized for production environments, with consistent output quality across batch operations—essential for teams managing large product catalogs or content libraries requiring AI photo editing at scale.
Key Considerations
- Complexity of Instructions: The model performs well with complex instructions, but clarity and specificity in prompts are crucial for optimal results.
- Input Quality: High-quality reference images can significantly enhance the model's performance in editing tasks.
- Resource Requirements: Running the model locally may require substantial computational resources.
- Quality vs Speed Trade-offs: Higher quality outputs may require longer processing times.
- Prompt Engineering Tips: Using detailed and specific prompts can help achieve desired outcomes.
Tips & Tricks
How to Use dreamomni2-edit on Eachlabs
Access dreamomni2-edit through Eachlabs via the interactive Playground or API integration. Provide your source image, text editing instructions, and optional reference images to guide the transformation. The model outputs high-resolution edited images in standard formats. Use the Eachlabs SDK for seamless integration into production workflows, batch processing, or custom applications requiring precise image-to-image AI editing at scale.
---END_CONTENT---Capabilities
- Multimodal Editing: Supports both text and image inputs for editing tasks.
- Abstract Concept Editing: Can manipulate abstract attributes like materials, textures, and hairstyles.
- Versatility: Offers a wide range of editing features, including style transfer and pose transfer.
- Quality of Outputs: Produces high-quality edited images comparable to commercial models.
- Technical Strengths: Integrates feature mixing and joint training with VLM for enhanced performance.
What Can I Use It For?
Use Cases for dreamomni2-edit
E-commerce product photography: Product teams can upload a product photo and provide a text prompt like "place this on a white marble countertop with soft morning light coming from the left" along with a reference image showing the desired aesthetic. dreamomni2-edit generates photorealistic composites that maintain accurate product appearance while matching the desired environment, eliminating expensive studio reshoot cycles.
Creative designers and art directors: Designers working on campaigns can use dreamomni2-edit to explore multiple visual directions without recreating assets from scratch. By providing a base image and text guidance ("make this more vibrant," "shift the color palette to cool tones") paired with style references, they iterate rapidly through design variations while maintaining semantic consistency in key elements.
Content creators and social media managers: Creators can adapt existing images for different platforms and seasons by editing backgrounds, adjusting lighting, or adding contextual elements. A travel blogger might transform a summer beach photo into an autumn scene by specifying "change to fall colors and add golden hour lighting" with a reference image, creating fresh content from existing assets.
Developers building image editing platforms: Teams developing AI image editor applications can integrate dreamomni2-edit via API to offer end-users multimodal editing capabilities. The model's support for both text and visual inputs enables developers to build interfaces where users can combine written instructions with mood boards or style references, creating more intuitive and powerful editing experiences than text-only alternatives.
Things to Be Aware Of
- Experimental Features: Some users report experimenting with novel applications of the model's abstract concept editing capabilities.
- Known Quirks: Users may encounter issues with pixel confusion if not using the index encoding scheme correctly.
- Performance Considerations: Requires significant computational resources, which can impact processing speed.
- Resource Requirements: Users need access to powerful hardware for optimal performance.
- Consistency Factors: Consistency in output quality can vary based on input quality and prompt clarity.
- Positive Feedback Themes: Users appreciate the model's ability to handle complex instructions and abstract concepts.
- Common Concerns: Some users express concerns about the model's resource requirements and potential limitations in handling very complex scenes.
Limitations
- Resource Intensity: The model requires substantial computational resources, which can limit accessibility for users without high-performance hardware.
- Complex Scene Handling: May struggle with very complex scenes or multiple abstract concepts simultaneously.
- Input Dependency: Performance is highly dependent on the quality and clarity of input prompts and reference images.
Pricing
Pricing Detail
This model runs at a cost of $0.050 per execution.
Pricing Type: Fixed
The cost remains the same regardless of which model you use or how long it runs. There are no variables affecting the price. It is a set, fixed amount per run, as the name suggests. This makes budgeting simple and predictable because you pay the same fee every time you execute the model.
Related AI Models
You can seamlessly integrate advanced AI capabilities into your applications without the hassle of managing complex infrastructure.
