MaSC: A Masked Similarity Metric for Evaluating Concept-Driven Generation

1Adam Mickiewicz University, 2Kiel University, 3ArtiCollect

Paper and arXiv links coming soon.

MaSC method diagram: a frozen SigLIP2 encoder produces both Concept Preservation and Prompt Following scores in a single forward pass, conditioned on a foreground concept mask.

MaSC pipeline overview. Reference and generated images are encoded once each by a frozen SigLIP2 SO400M-NaFlex vision tower, producing patch-token grids R, G ∈ ℝN×D. MaSC consumes provided foreground concept masks, denoted MRF, MGF ∈ {0, 1}H×W. Two branches share the single forward pass. Branch 1 — Concept Preservation (masked-maxcos): for each patch covered by the reference foreground mask (MRF), take its maximum cosine similarity against all output patches in G, then average over the foreground-reference patches. Branch 2 — Prompt Following (Text-Background Alignment): the trained SigLIP2 attention pooler is run with foreground patches of G masked from the attention computation, producing a background-only embedding. This is compared via cosine similarity to the text embedding etext of the prompt with the subject noun stripped.

Abstract

Evaluating single-concept personalization in text-to-image diffusion has seen two categories of quantitative metrics: Concept Preservation (CP) measures identity fidelity to a reference while Prompt Following (PF) measures whether the generated scene matches the prompt. Personalization papers have commonly computed these signals using three separate backbones: CLIP-I and DINO for CP, CLIP-T for PF. In this paper, we show that existing metrics fall short of correlating with human perception because they attend to the image as a whole, instead of distinguishing the concept subject from the background. This distinction is important from a human perception point of view as the concept subject in the output image should be very similar to the input concept (CP), whereas the output background should adhere closely to the text prompt (PF). To improve personalization evaluation in this way, we introduce MaSC, a unified metric that attends to CP and PF by differentiating concept subject regions. Specifically, given an externally provided foreground concept mask, MaSC computes the CP and PF scores from a single forward pass of a frozen SigLIP2 encoder per image.

On DreamBench++ human ratings, MaSC reaches Krippendorff α = +0.471 on CP — beating every non-LLM baseline tested (DINOv3, DreamSim, AM-RADIO, DIFT-SDXL, DINO-I, CLIP-I) and GPT-4V, while sitting within Δα = 0.028 of GPT-4o. To distinguish human perception bias we also evaluate our method, as well as common SOTA baselines, on ORIDa, a real-photo benchmark of identity preservation across physical environments. In this experiment MaSC reaches AUC = 0.992 almost perfectly identifying concept subjects. The PF score, obtained by MaSC without a second encoder forward pass, beats the CLIP-T baseline shipped with DreamBench++. In summary, our comprehensive evaluations demonstrate that MaSC establishes a new state-of-the-art for non-LLM concept preservation, while providing an efficient, unified standard for personalization evaluation. We release MaSC as a pip-installable Python package, alongside independent reproductions of every comparator's published numbers.

BibTeX

@misc{bartkowiak2026masc,
  title        = {{MaSC}: A Masked Similarity Metric for Evaluating Concept-Driven Generation},
  author       = {Bartkowiak, Patryk and Petersen, Lennart and Kotrys, Bartosz and Pirk, S{\"o}ren and Pa{\l}ubicki, Wojtek},
  year         = {2026},
  note         = {Preprint},
  howpublished = {\url{https://masc-metric.github.io/}}
}