Designed an AI platform that reduced catalog editing time by 60% (15→6 hrs per 100-SKU catalog)
TEAM
1 PM, 4 SDE, 1 Designer
ROLE
Product Design + AI Workflows
TIMELINE
April–August 2025 (20 weeks)
Context
What is Photogenix?
Many E-commerce brands were losing sales due to costly photoshoots, slow editing cycles, and inconsistent imagery.
To address this, I helped design Photogenix, an AI-powered product imaging platform that helps fashion and retail brands generate production-ready visuals at scale. Initially spun out from a micro-feature in Catalogix (just with background change + marketplace resizing), and quickly evolved into a standalone platform with a broader set of capabilities.
I was involved from 0→1 — identifying pain points, defining requirements, and shaping the solution from the ground up.
Impact So Far
Efficiency
↓60%
processing time
Cost
↓₹100
per image (~65%)
Scale
100k+
images processed so far
The platform made catalog creation faster, cheaper, and more consistent, helping brands deliver high-quality visuals with a fraction of the time and cost.
Opportunity
Making catalog imaging a growth lever
Photoshoots are slow, expensive, and brittle. They block catalog velocity, force compromises on quality, and prevent brands (especially smaller ones) from staying visually relevant to fast-moving trends.
Why this matters
Revenue impact
Catalog visuals directly drive conversion, delays in refreshing imagery mean missed seasonal demand and lost revenue
Operational bottlenecks
Workflows depend on expensive, hard-to-scale resources (studios, models, photographers), creating scheduling conflicts and revision cycles
Competitive disadvantage
Smaller brands can't afford frequent shoots, so they recycle outdated photos that don't reflect current trends, impacting discovery and conversion
Research & Insights
What We Heard and Learned (Selected Insights)
Through many in-depth user interviews and spending time with the cataloging team, we uncovered several key insights:
Operational pain
Frequent scheduling conflicts, last-minute styling/model changes, and endless revision cycles
Budget constraints
Tight budgets force compromises on quality or scale — smaller brands can't match marketplace leaders' imagery
Image freshness gap
A majority of merchants reported using photos 6–12 months old because they couldn't turnaround new shoots quickly.
AI usability gap
Off-the-shelf AI tools tried by merchants "look fake" or require complex prompting. Users want easy-to-use, simple controls
Key Insight: Users needed affordable, time efficient solution that also gives them creative control without technical complexity. They wanted to focus on brand decisions, not prompt engineering.
Amazon wants one size, Myntra wants another size. By the time we edit all photos, two days gone minimum.
We need a Y2K street-style look for our next collection — finding the right model and location will push our launch by at least a week.
Customers want to see how shirt looks in office setup, casual setup - but doing multiple location shoots is not feasible for us.
Tried 3-4 AI tools. Either the output looks fake or they ask us to write some complicated prompt. Who has time to learn all this technical stuff?
Scoping
Problem Definition
How might we enable non-technical merchants to create production-ready, contextual product imagery that matches their brand standards without complex AI prompting or expensive photoshoots?
Design Principles (Derived from Research)
Cognitive load reduction
Hide AI complexity behind familiar metaphors
Predictable outcomes
Users should know what they'll get before generating
Flexible and simple interface
Support both simple and advanced use cases in one flow
Production-ready visuals
Output should require minimal post-processing
Product hypotheses:
If we provide preset driven, single-click contextual variants, merchants will refresh images more often because it's faster and cheaper than scheduling shoots.
If we hide prompt complexity behind curated presets, adoption and satisfaction among non-technical users will rise.
If we integrate automatic marketplace resizing/export profiles, post-processing time will drop significantly, removing a major friction point.
Solution Development
A Simple, Predictable Workflow
Based on the derived design principles, we designed Photogenix around a simple 3-step flow:
This approach minimized friction for merchants who were frustrated with complex, text-based prompting.
A merchant uploads an image, and Photogenix instantly shows tailored options based on the use case — on-model generation, mannequin-to-model transformation, product background change, or flat-lay integration. The user can make quick selections for model, background, footwear, or any specific visual element they want to modify or keep.
Through early feedback, we discovered that not every merchant needed the full workflow. Some only wanted to replace a background, while others just needed a new model. Instead of fragmenting the experience with multiple tools, we introduced a “Keep Original” option within the same flow. This gave users the flexibility to change the background, the model, or both — without context-switching or juggling separate modules.
For added creative control, merchants can also upload or describe a model, background, or other desired attributes within each section, ensuring complete customization while keeping the interface intuitive.
A Simple, Predictable Workflow
Since Photogenix supports multiple content types — from static product photos to generative videos — we applied a unified, card-based design system across all workflows. The preset architecture keeps the visual logic consistent, while modular controls adapt dynamically to each use case. This system ensures a predictable, scalable experience whether the user is editing a product, mannequin, or video frame.
Change the background, model, and footwear effortlessly — all within one unified flow.
Turn mannequin or ghost mannequin shots into realistic on-model images. Retains silhouette, drape, and color accuracy.
Clean product cutouts with smart background replacement, shadows, and reflections. Export marketplace-ready sizes in one go.
Go beyond basics — style your scene with custom backgrounds, props, and accessories for complete creative control.
Compose ad-ready visuals using brand presets, text overlays, and aspect ratios for paid and organic placements. Batch render variants.
Create short product clips or animate stills into smooth loops — control camera motion, duration, and quality using multiple state-of-the-art models in one seamless workflow.
Scaling for Real Workflows
As adoption grew, merchants wanted to process entire catalogs at once instead of running image generation SKU by SKU.
Bulk and Zip File Upload
Designed for large-scale catalog operations.
- Users upload a zip file with multiple folders, each representing a product or category.
- Each folder can have custom settings (e.g., office look for shirts, street look for sneakers, minimal studio look for accessories).
- The system processes everything asynchronously in the background. Users are free to continue exploring the platform, and they get notified once assets are ready.
With the bulk feature, hundreds of product images can be generated overnight, without constant user input.
Feedback implementation
Addressing AI Output Limitations Through UX
While Photogenix produced high-quality, production-ready images, generative AI models still had edge-case failures:
- Artifacts in complex areas (garments with fine textures, accessories not rendering correctly)
- Anatomical distortions (hands/fingers, a common SDXL limitation)
- Background blending issues (shadows, edges not harmonized)
- Inconsistent lighting or mismatched expressions
From looking at Posthog sessions and feedback sessions, ~30% of generated outputs showed at least one of these issues (n=500 images, July 2025) — not enough to break the product, but significant enough to require an efficient, user-friendly correction path.
Instead of forcing users to regenerate repeatedly, we introduced a retouch toolkit inside the same "Renders" screen — so fixes could be made quickly without leaving the workflow.
Key tools we added
Users can mask an area (e.g., sleeve, logo, accessory) and enter a short prompt to re-generate just that section.
When the auto-mask mis-detected areas, users could manually adjust the mask to ensure precise retouching.
A one-click adjustment harmonized lighting across subject and background, with simple approve/reject.
A lightweight control to adjust facial expressions without re-shooting or re-generating the full image.
Users felt in control rather than at the mercy of AI. Fixes took an average of 45 seconds compared to 8-15 minutes for regeneration or Photoshop edits. Despite AI’s limitations, professional output quality remained consistent.
Iterations Based on Feedback
We continuously refined Photogenix based on insights from user interviews and PostHog session analysis, using observed friction points to improve clarity, scalability, and speed across the experience.
Through these cycles, we also realized that feature requests would keep evolving — merchants often asked for new controls or editing tools.
So, we made a strategic decision to redesign the interface with scalability in mind, ensuring that new capabilities could be added seamlessly without disrupting the core workflow.
Homepage & Generation Screen
In our early launch, users selected their subject type (model, mannequin, or product) from a single modal pop-up. While simple, this approach became confusing as we expanded to new capabilities like flat-lay to model, video, and ad generation. It wasn’t scalable.
We replaced the modal with dedicated feature cards, each visually representing its use case. This made discovery effortless and also solved a key user pain: merchants often uploaded poor-quality input images, not realizing how it affected results. Each card now displays examples of good vs. poor inputs, helping users choose the right type of image before generation.
For the generation workflow, the previous sidebar layout made the interface visually dense and harder to extend. We consolidated all options into a single, modular panel that opens in context. This simplified the visual hierarchy and made it easier to scale as we added advanced controls like “Additional Prompt,” “Retouch,” and “Lighting.”
Edit Mask
The original mask editor looked intuitive, but session recordings revealed repeated failed attempts — users couldn’t tell exactly what was being masked.
We redesigned it to show the editable mask in real-time, directly on the image. This gave users instant visual feedback and far greater precision during edits.
Renders Screen
Initially, we displayed generation results in a modal popup showing the original vs. generated image, credits used, and editing options like Retouch and Relight. As features expanded, this layout became cramped and hard to navigate.
We transitioned to a full-screen render view, adding a right-hand sidebar for all editing tools and detailed information. This not only improved readability but also allowed us to compare image variations side by side to ensure visual consistency.
Continuous Micro-Improvement
Beyond these larger redesigns, we shipped continuous refinements based on usage data — improving tool discoverability, clarifying actions, and removing unnecessary steps to keep the experience simple and predictable as new capabilities were introduced.
Beyond Design — Contributions to AI Workflows
Though my primary role was product design, I also contributed directly to the AI workflow exploration. I independently learned ComfyUI to better understand how workflows were constructed.
- Model-to-Model & Mannequin-to-Model workflows: Built early prototypes integrating Depth and Canny ControlNets to preserve pose and structure while maintaining stylistic flexibility. These prototypes provided a working foundation for our generative fashion pipeline.
- Continuous improvements: Iterated across multiple model updates, fine-tuning prompt conditioning, sampler settings, and ControlNet weights to improve realism, lighting consistency, and fabric detail.
- Bridging design & AI: This effort helped me translate design intent into technical workflows, ensuring that user-facing features (like mannequin-to-model conversion) were feasible and aligned with real model capabilities
Impact & Outcomes
Photogenix meaningfully improved catalog workflows for merchants across fashion and retail:
Efficiency
↓60%
processing time
Cost
↓₹100
per image (~65%)
Scale
100k+
images processed so far
TIME-TO-PUBLISH
Same-day
vs 2.2 days before
CTR UPLIFT
+6–10%
on SKUs with refreshed imagery
Learnings & Reflections
What I Learned About AI Product Design
Designing Photogenix taught me that impactful AI products are less about "magic" and more about control, trust, and recovery. The best outcomes came from combining user empathy, technical curiosity, and awareness of business constraints. Users didn't need perfect outputs — they needed confidence the system worked with them, plus simple ways to fix mistakes.
Design craft evolved through:
Design craft evolved through:
- Systematic user research and behavioral analysis
- Data-driven iteration and validation
- Cross-disciplinary learning (AI workflows, business metrics)
- Building trust through transparency and user control
Photogenix was a true 0→1 journey — from identifying pain points to shaping the end-to-end product experience. The work showed me the value of combining user-centered design with technical curiosity, and reinforced that impactful product design is not just about polished screens, but about making workflows faster, cheaper, and more reliable at scale.
Photogenix V1 Case Study
Explore the first version of PhotoGenix and see how our design and functionality have grown.
Click here to view