link href="https://fonts.googleapis.com/css?family=Kaushan+Script:regular" rel="stylesheet" />


CARE-Edit
Condition- Aware Routing of Experts
for Contextual Image Editing


Anonymous Authors
for review

We propose CARE-Edit , a contextual image‑editing framework using Condition- Aware Routing of Experts . It selects expert branches based on editing conditions, enabling precise and robust edits in a unified model.


Abstract


Unified diffusion editors often rely on a fixed, shared backbone for diverse tasks, suffering from task interference and poor adaptation to heterogeneous demands (e.g., local vs global, semantic vs photometric). Prevalent ControlNet and OmniControl variants combine multiple conditioning signals (e.g., text, mask, reference) via static concatenation or additive adapters that cannot dynamically prioritize or suppress conflicting modalities, leading to artifacts like color bleeding across mask boundaries, identity or style drift, and unpredictable behavior.

CARE-Edit Abstract To address this, we propose Condition-Aware Routing of Experts (CARE-Edit) that aligns model computation with specific editing competencies. At its core, a lightweight latent-attention router assigns encoded diffusion tokens to four specialized experts—Text, Mask, Reference, and Base—based on multi-modal conditions and diffusion timesteps: (i) a Mask Repaint module first refines coarse user-defined masks for precise spatial guidance; (ii) the router applies sparse top-K selection to dynamically allocate computation to the most relevant experts; (iii) a Latent Mixture module subsequently fuses expert outputs, coherently integrating semantic, spatial, and stylistic information to base images.


How does it work?


CARE-Edit overview

CARE-Edit introduces condition‑aware specialized experts within the frozen DiT backbone. Given multimodal conditions, inputs are tokenized and projected to heterogeneous expert branches. The router assigns confidence scores and selects the top‑K experts to process each token. Expert outputs are normalized, modulated, and fused through the Latent Mixture module, yielding denoised representations h' refined by Mask Repaint. Only lightweight adapters, the router, and fusion layers are trainable. This enables CARE-Edit to dynamically allocate computation, mitigates conflicts between heterogeneous conditions (e.g., text vs mask) and enables high-fidelity, coherent edits.


Generalization Capabilities


Contextual Image Editing

Style Transfer

Object Removal

Object Addition & Replacement




Comparison with State-of-the-Art Methods


Single condition comparison

Multi-condition comparison

Disclaimer


This project page is prepared for academic presentation of the CARE-Edit framework. All visual examples are for research and illustration purposes only. If any content is considered inappropriate or infringes rights, please contact the authors, and we will promptly address the issue. This offline HTML page is adapted from the UNO project webpage. We sincerely thank the UNO authors for open-sourcing their project page design.



BibTex (placeholder)


@article{anonymous2025careedit,
  title={CARE-Edit: Condition-Aware Routing of Experts for Contextual Image Editing},
  author={Anonymous Authors},
  journal={Under review},
  year={2025}
}



Template adapted from UNO Github Pages.