1 / 15
EvoMUSART 2026

Prompt and Circumstances

Evaluating the Efficacy of
Human Prompt Inference
in AI-Generated Art

Khoi Trinh, Scott Seidenberger  ·  University of Oklahoma
Joseph Spracklen, Raveen Wijewickrama, Murtuza Jadliwala  ·  UT San Antonio
Bimal Viswanath  ·  Virginia Tech   |   Anindya Maiti  ·  University of Oklahoma
📅 EvoMUSART 2026  ·  Toulouse, France  ·  April 2026
2 / 15

The Prompt Economy

Prompt Marketplaces

Platforms like PromptBase, PromptHero, and CivitAI let creators buy, sell, and share text prompts for AI image generation.

Prompts as IP

Sellers assert intellectual property rights over their prompts, claiming them as proprietary creative assets. Platforms keep the actual prompt hidden until purchase.

The Vulnerability

Every marketplace listing shows sample images. Can someone reverse-engineer the secret prompt just by looking at those images?

PromptBase marketplace listing example (Fig. 1)
Fig. 1 — Example PromptBase listing: sample images shown, prompt hidden until purchase.
Key question: If the hidden prompt can be recovered from the sample images, does it still constitute meaningful intellectual property?
3 / 15

What is Prompt Inference?

✍️ Original Prompt Secret text used to generate target image
🤖 txt2img Model MidJourney, SDXL, DreamShaper…
🖼️ Published Image Visible on the marketplace listing
🔍 Inferred Prompt Human (or AI) reverse-engineers the text
📊 Compare Do generated images look the same?

Prompt Anatomy

Prompts have a subject (what it depicts: cat, robot, astronaut…) and one or more modifiers (style, lighting, mood, color palette, medium…)

Even small modifier changes radically alter the image — capturing stylistic intent from visuals alone is deeply ambiguous.

4 / 15

Research Questions & Hypotheses

RQ1 How accurately can humans infer the original text prompt from an AI-generated image?
RQ2 How do prompt length, complexity, modifier ambiguity, and model choice affect inference performance?
RQ3 Can large language models (LLMs) improve human prompt inference through collaborative, combined prompt construction?
H1 Humans infer prompts with moderate success — recognizing subjects, but struggling with nuanced stylistic modifiers.
H2 Inference success correlates with prompt characteristics (length, adjectives, clause counts) and the generation model.
H3 Combining human-inferred prompts with AI suggestions (CLIP Interrogator + GPT-4) yields higher similarity scores than human inference alone.
5 / 15

Study Overview

230
Participants
(campus + MTurk)
200
Total prompts
(100 controlled + 100 uncontrolled)
2,286
Valid prompt responses
retained after filtering

Part I — Uncontrolled Dataset

100 real-world prompts sampled from PromptHero. Wide stylistic variety. Participants wrote free-form prompts describing five AI-generated images they were shown. Open-ended inference task.

Part II — Controlled Dataset

100 prompts built from scraped Lexica data. Fixed subjects (man, woman, astronaut, cat, robot) + 121 stylistic modifiers from Midjourney's keyword list. Ground truth known; enables direct comparison.

6 / 15

Datasets & Models

Four txt2img Models

  • MidJourney v5 — closed-source, Discord interface; highly stylized
  • Stable Diffusion XL (SDXL) — open-source transformer-based
  • DreamShaper XL — SDXL fine-tune for photo-realistic fantasy imagery
  • Realistic Vision v5 — SD 1.5 fine-tune for photo-realism

CFG Scale: 5 · Sampling Steps: 40 · Euler sampler
MidJourney used default model settings (parameters not user-controllable)

Participant Profile (n=230)

  • Most aged 18–44; 59% male, 38% female, 2% other
  • Moderate familiarity with generative AI: ~⅓ were only "slightly familiar" with image generation tools
  • Participants saw 5 images each from controlled + uncontrolled sets and typed their best-guess prompts

Responses filtered: excluded entries missing subject or modifier, containing random/blank text, or non-English

Example images shown to participants during the study
7 / 15

Evaluation: Image Similarity Metrics

We measure inference quality by generating images from inferred prompts and comparing them to original-prompt images using three metrics:

ImageHash

Perceptual hash via Hamming distance. Score 0–64 (lower = more similar). Captures shallow visual cues.

LPIPS

Learned Perceptual Image Patch Similarity in deep feature space. Score 0–1 (lower = more similar). Reflects structure, texture, color composition.

CLIP Score

Cosine similarity of CLIP (ViT-L/14, ViT-B/32) embeddings. Score 0–1 (higher = more aligned). Measures semantic text–image alignment.

Primary metrics used: LPIPS and CLIP (B32 + L14) — they better capture perceptual and semantic fidelity than shallow hash matching.
8 / 15

A Novel Evaluation Framework: Distribution-Level Inference

The Problem with Single-Image Comparison

txt2img models are stochastic — the same prompt produces different images each run. A single lucky match doesn't prove prompt equivalence.

Our Approach: Distributional Equivalence

For each prompt, generate 200 reference images (original prompt) and 50 inferred images (participant prompt). Treat each as a sample from a probability distribution.

Kolmogorov–Smirnov Test

Two-sample KS test checks whether the two similarity-score distributions are statistically indistinguishable (p > 0.05 = a "hit").

Fig. 6 — Same prompt (a,b) → overlapping distributions (f); different prompt (d) → diverging distribution (h). A hit = distributions are statistically indistinguishable (p > 0.05).

9 / 15

Human–AI Combined Prompt Inference

👤 Human Prompt Participant's written inference
+
🔬 CLIP Interrogator AI-generated caption aligned to CLIP embeddings
🤝 GPT-4 Merge Instructed to synthesize a single ≤25-word prompt
📊 Combined Prompt Evaluated vs. human-only baseline

Example Merge

Human: "man dressed in steampunk in a steampunk factory"
CLIP Interrogator: "a man in a steampunk suit and top hat standing in front of a giant clock with gears, Bastien L. Deharme, fantasy art"
GPT-4 output: "a man in a steampunk suit and top hat stands in a factory surrounded by gears and clocks, embodying a fantasy art portrait."

LLM merging was chosen over simple concatenation to preserve linguistic coherence, balance human and AI contributions, and avoid keyword repetition that could bias image generation.

10 / 15

Results — RQ1: Human Prompt Inference Performance

Metric Controlled Hit Rate Uncontrolled Hit Rate
ImageHash 53.29% 53.72%
LPIPS 22.89% 9.77%
CLIP B32 7.28% 7.56%
CLIP L14 7.05% 7.21%
⚠ ImageHash only captures shallow pixel structure — can produce false positives — excluded from final analysis
LPIPS & CLIP: stricter perceptual/semantic metrics → more meaningful ground truth.
H1 — Partial Participants reliably recover subjects but struggle with modifiers. Performance drops substantially from controlled (structured) to uncontrolled (open-ended) prompts, confirming that stylistic nuance is the bottleneck.
Key insight: CLIP hit rates (~7%) reveal that semantic alignment is very hard to achieve through visual inference alone, even in controlled settings. Humans can roughly identify "what" but not "how" the image was styled.

LPIPS gap between controlled (22.89%) and uncontrolled (9.77%) confirms that open-ended prompts from the wild are far harder to reverse-engineer than structured laboratory prompts.

11 / 15

Results — RQ2: Linguistic Features & Model Effects

Linguistic Features (CLIP & LPIPS)

  • Longer prompts only weakly correlate with higher hit rates
  • CLIP mean correlates positively with hit rates
  • LPIPS mean correlates negatively with hit rates
H2 — Partial Linguistic richness has modest, consistent effects. Neither prompt complexity nor model choice is a strong predictor of inference success.

Model Choice Effects

DreamShaper XL — strongest positive relation with success (controlled r= +0.15).

Realistic Vision v5 — higher variability (controlled r= +0.14 for LPIPS variance), weaker inference success.

SDXL — largely neutral. Trends consistent across controlled and uncontrolled sets.

12 / 15

Results — RQ3: Does Human–AI Collaboration Help?

Metric Human Only Human–AI Combined Change
ImageHash 53.29% 60.62% ↑ +7.3pp
LPIPS 22.89% 18.59% ↓ −4.3pp
CLIP B32 7.28% 5.89% ↓ −1.4pp
CLIP L14 7.05% 5.77% ↓ −1.3pp
H3 — Rejected Human–AI merging did not improve LPIPS or CLIP performance. It actually reduced them, while only improving shallow ImageHash similarity.
Naïve merging strategies often diluted semantic coherence, resulting in lower similarity scores under meaningful evaluation metrics.

ImageHash improvement (+7pp) reflects better pixel-level surface similarity, not meaningful perceptual alignment. This is a false positive — and further confirms why we exclude ImageHash from core analysis.

13 / 15

Discussion: Key Takeaways

What Humans Can Do

Broad subject matter is recoverable from images. Participants reliably identify the main theme, object, or scene — especially in controlled settings with common subjects.

What Remains Hard

Precise stylistic modifiers — lighting, texture, color palette, artistic medium, mood — are extremely difficult to reconstruct. CLIP hit rates of ~7% confirm this ceiling.

Metric Selection Matters

ImageHash inflates success by capturing only shallow pixel patterns. LPIPS and CLIP are more honest evaluators and reveal genuinely low inference success for unconstrained prompts.

Implications for Prompt IP

Results suggest that prompts are relatively resilient to reverse engineering under current human + AI inference techniques.

Human–AI Collaboration Gap

Effective human-AI co-creation for prompt inference remains unsolved. Future methods should optimize for LPIPS/CLIP alignment, not shallow surface similarity.

User Background Doesn't Matter Much

AI familiarity and arts background had limited impact on success rates — task and system-level constraints dominate over user expertise.

14 / 15

Conclusion

  • Humans can infer what an image depicts, but not how it was styled — subjects yes, modifiers no
  • LLM-augmented inference (CLIP Interrogator + GPT-4 merging) does not outperform human-only prompting on meaningful metrics
  • Prompt-based IP appears relatively resilient to reverse engineering under current techniques
  • Raises key questions for web platforms around content provenance, authorship, and secure prompt design
~7%
CLIP semantic hit rate (human inference)
Human–AI merging reduced LPIPS & CLIP vs. human-only
Novel KS-based distributional evaluation framework
15 / 15

Q&A

Full Paper Pre-print

arXiv pre-print QR / preview
🙋
Questions?
Khoi Trinh · University of Oklahoma
khoitrinh@ou.edu
Paper: Prompt and Circumstances: Evaluating the Efficacy of Human Prompt Inference in AI-Generated Art
EvoMUSART 2026
1 / 15