1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
const axios = require('axios');
const fs = require('fs');
const path = require('path');
async function toB64(imgPath) {
const data = fs.readFileSync(path.resolve(imgPath));
return Buffer.from(data).toString('base64');
}
const api_key = "YOUR API-KEY";
const url = "https://api.segmind.com/v1/qwen-image-edit-plus-product-photography";
const data = {
"prompt": "Convert the white background image to a scene and place the sofa in the picture in a modern living room",
"image_1": "toB64('https://segmind-resources.s3.amazonaws.com/input/8e3e4da5-f25b-49df-ba41-048b279a9f62-10_ELlaximag.webp')",
"image_2": "toB64('')",
"image_3": "toB64('')",
"lora": "product_photography",
"aspect_ratio": "match_input_image",
"seed": 87568756,
"image_format": "webp",
"quality": 95,
"base64": false
};
(async function() {
try {
const response = await axios.post(url, data, { headers: { 'x-api-key': api_key } });
console.log(response.data);
} catch (error) {
console.error('Error:', error.response.data);
}
})();Text prompt describing the desired image edit or generation
Primary input image (URL or base64)
Secondary input image (URL or base64)
Tertiary input image (URL or base64)
Pre-configured LoRA model to apply
Allowed values:
Additional LoRA model URL. Public direct url or huggingface url pointing to the lora file
Additional LoRA model URL. Public direct url or huggingface url pointing to the lora file
Output image aspect ratio
Allowed values:
Random seed for reproducibility. Use -1 for random
min : -1,
max : 2147483647
Output image format
Allowed values:
Output image quality (1-100)
min : 1,
max : 100
Return image as base64 encoded string
To keep track of your credit usage, you can inspect the response headers of each API call. The x-remaining-credits property will indicate the number of remaining credits in your account. Ensure you monitor this value to avoid any disruptions in your API usage.
Qwen-Image-Edit-2509-White_to_Scene is a specialized image-to-image AI model fine-tuned from Qwen/Qwen-Image-Edit-2509 for transforming white background images into rich, contextual scenes. This model excels at seamlessly placing subjects—like products, vehicles, or people—into realistic outdoor or indoor environments. Built on the Diffusers library and leveraging advanced image fusion techniques, it enables creators to generate professional-quality edits without complex compositing workflows. Whether you're building an e-commerce product visualizer or creating marketing assets, this model delivers high-fidelity scene generation with natural lighting and environmental context.
E-commerce and Product Photography: Transform product shots on white backgrounds into lifestyle scenes—place furniture in living rooms, cars on mountain roads, or watches in luxury settings without expensive photoshoots.
Marketing and Advertising: Generate multiple environmental variations of the same subject for A/B testing, seasonal campaigns, or regional market adaptation.
Real Estate and Automotive: Visualize vehicles in different settings or stage empty spaces with realistic context for listings and promotional materials.
Creative Design and Mockups: Rapidly prototype visual concepts by placing design elements into realistic scenes for client presentations or concept validation.
Be Specific About Environment Details: Instead of "outdoor scene," write "sunny afternoon in a modern urban plaza with trees and pedestrians" for precise results.
Layer Your Descriptions: Structure prompts with subject placement first, then environment, then lighting: "Place the red sedan in a mountain landscape during golden hour with soft shadows."
Leverage the Product Photography LoRA: When editing commercial items, the built-in LoRA automatically enhances lighting, reflections, and material rendering—specify desired qualities like "natural glow" or "studio lighting."
Use Seed Values Strategically: Set a specific seed (not -1) when iterating on prompts to isolate the effect of prompt changes. Switch to -1 for exploring creative variations.
Aspect Ratio Considerations: Match your output ratio to the target platform—use 1:1 for Instagram posts, 16:9 for YouTube thumbnails, or "match_input_image" to preserve original dimensions.
Quality vs. File Size: For web delivery, quality settings between 85-95 with WebP format offer the best balance of visual fidelity and loading speed.
Is Qwen-Image-Edit-2509-White_to_Scene open-source?
This is a fine-tuned version of the Qwen-Image-Edit-2509 base model. Check the original model's licensing on ModelScope or Hugging Face for commercial usage terms.
How is this different from general image-to-image models?
Unlike generic editors, this model is specifically trained for white-background-to-scene transformation. It understands how to maintain subject integrity while generating coherent environmental context, lighting, and shadows—eliminating common compositing artifacts.
What parameters should I tweak for best results?
Start with a descriptive prompt and the default product_photography LoRA. Adjust aspect_ratio to match your output needs. Use a fixed seed during prompt refinement, then remove it for final variations. Set quality to 95 for print or master assets.
Can I use custom LoRA models?
Yes—the model accepts two additional LoRA URLs (lora_2_url, lora_3_url) for style customization. This enables you to layer brand-specific aesthetics or artistic filters on top of the base scene generation.
Why use multiple images instead of just one?
Multi-image inputs enable advanced workflows like subject swapping, element blending, or before/after comparisons within a single generation. This is particularly useful for product variations or compositional experiments.
What's the recommended workflow for production use?
Upload a clean white-background image, write a detailed environment prompt, select your aspect ratio, and set a seed. Review the output, refine your prompt based on results, then generate final variations with seed set to -1 for creative diversity while maintaining the refined prompt structure.