Search AI-Generated Product Photos by Meaning, Not File Names

Once a fashion team starts using AI for production imagery, the bottleneck changes. The hard part is no longer only creating one good product photo. It is finding the right one again after the team has generated hundreds of studio shots, model variants, campaign concepts, close-ups, edits, upscales, and video frames.
Customers asked us for this because collaborative asset libraries get messy quickly. One teammate generates a useful result, another teammate wants to edit it, upscale it, or use it as the input for a new shoot, and suddenly the team is scrolling through thousands of generated results.
That is why Uwear now supports semantic search for generated product photos. Instead of relying only on file names, tags, items, dates, owners, or generation IDs, teams can search by what they remember seeing: "ski mountains", "little girl pink", "denim close-up", or "white background packshot".
Short version: Uwear is turning generated fashion images into a searchable team asset library. New results can be embedded asynchronously, stored in pgvector, and retrieved by semantic meaning, so teams can find assets by what they see instead of where someone filed them.
In the demo, searches like "ski mountains" and "little girl pink" return the visual result the user remembers, even when a normal filter would not describe it.
Why Product Photo Libraries Break
Traditional ecommerce asset libraries depend on structured metadata: SKU, collection, product name, shoot date, folder, color, and tags. That still matters, and Uwear keeps those filters important. But generated content creates a new retrieval problem because the most useful description is often visual, not administrative.
A merchandiser might remember "the version with softer studio lighting". A creative lead might remember "the man skiing in the mountains". A performance marketer might need "the little girl in a pink sweatshirt". Those are not always tags someone planned in advance.
Searches this is designed for
- -"ski mountains"
- -"little girl pink"
- -"red dress on a city street"
- -"close-up denim texture"
- -"minimal white-background packshot"
What Semantic Search Means Here
Semantic search compares meaning, not exact words. For generated fashion images, that means the system can represent an image as an embedding, represent a text query as an embedding, and return generated results whose visual content is close to the query.
It is not a replacement for product data. It is a second retrieval layer that works especially well when the search starts with a visual memory: garment type, color, texture, scene, pose, angle, background, or mood.
| Search type | Best for | Example |
|---|---|---|
| Metadata filters | Known fields like product, date, tag, owner, kind, or batch. | All image results for one jacket from last week. |
| Semantic search | Visual concepts that are hard to tag exhaustively. | Images that feel like "soft studio light, sage green jacket". |
| Together | Production review, variant discovery, and creative reuse. | Semantic search inside one collection, one garment, or one content type. |
How Uwear Indexes Generated Images
The important product choice is that indexing happens asynchronously. When a generation finishes, Uwear can queue the available results for embedding work instead of making the user wait for search infrastructure before seeing the output.
Generation finishes -> available image results are queued for embedding -> the embedding worker reads the image or thumbnail -> Gemini creates a 768-dimensional image embedding -> Uwear stores it in Postgres with pgvector -> search queries are embedded and ranked by cosine distance
What changed under the hood
- -Embedding storage: generated result embeddings are stored in Postgres using pgvector.
- -Gemini embedding worker: an async worker embeds generated result images without blocking generation completion.
- -Similarity ranking: semantic results are ordered by embedding distance, then recent results.
- -Weak-match filtering: Uwear applies a maximum distance threshold so vague matches are filtered out.
- -Backfill awareness: older historical results may need to be indexed before they appear in semantic search.
How Teams Use It
Semantic search is most useful when Uwear becomes a daily production library shared by a team. A user can find a colleague's result, open it, edit it, upscale it, use it as a starting image for a new shoot, or turn it into video without needing to know who generated it or what filters would surface it.
Regular filters still help when the team knows the product, date, type, tag, or owner. Semantic search helps when the team only remembers the image: the scene, garment, pose, color, model, background, or mood.
Workflow examples
- -Collaborative reuse: find a teammate's generated result without asking them to send the link or generation ID.
- -New shoot inputs: locate an existing result and use it as the source image for another photoshoot, edit, or video.
- -Creative review: find the strongest concepts from a large batch without opening every generation.
- -Campaign reuse: search past generations for visual directions that can be reused in a new launch.
- -Catalog QA: locate similar poses, backgrounds, or close-ups across a whole collection.
Available in Studio and API Workflows
In Uwear Studio, semantic search appears in generated-result views where teams already review and select assets. For developers and agents, the generated-results API can accept a semantic query alongside the usual filters and pagination.
GET /generation-results?semantic_query=close-up%20denim%20texture Authorization: Bearer YOUR_API_KEY
The bigger idea is simple: as fashion brands generate more AI product content, search has to become visual and intent-aware. Metadata helps you organize the library. Semantic search helps you remember what is inside it.
Build a searchable AI product photo library
Use Uwear to generate, review, find, edit, upscale, and reuse fashion visuals across your product and campaign workflows.