The Unreasonable Effectiveness of
Text Embedding Interpolation
for Continuous Image Steering

Reve

Continuous Steering Examples

Steering example 1, frame 1 Steering example 1, frame 2 Steering example 1, frame 3 Steering example 1, frame 4 Steering example 1, frame 5 Steering example 1, frame 6

Zoom Out

Steering example 2, frame 1 Steering example 2, frame 2 Steering example 2, frame 3 Steering example 2, frame 4 Steering example 2, frame 5 Steering example 2, frame 6

Cartoon Style

Steering example 3, frame 1 Steering example 3, frame 2 Steering example 3, frame 3 Steering example 3, frame 4 Steering example 3, frame 5 Steering example 3, frame 6

Rotten

Steering example 4, frame 1 Steering example 4, frame 2 Steering example 4, frame 3 Steering example 4, frame 4 Steering example 4, frame 5 Steering example 4, frame 6

Dog -> Bear

Steering example 5, frame 1 Steering example 5, frame 2 Steering example 5, frame 3 Steering example 5, frame 4 Steering example 5, frame 5 Steering example 5, frame 6

Photorealism

Steering example 6, frame 1 Steering example 6, frame 4 Steering example 6, frame 5 Steering example 6, frame 6

Size

Abstract

We present a training-free framework for continuous and controllable image editing at test time for text-conditioned generative models. In contrast to prior approaches that rely on additional training or manual user intervention, we find that a simple steering in the text-embedding space is sufficient to produce smooth edit control. Given a target concept (e.g., enhancing photorealism or changing facial expression), we use a large language model to automatically construct a small set of debiased contrastive prompt pairs, from which we compute a steering vector in the generator’s text-encoder space. We then add this vector directly to the input prompt representation to control generation along the desired semantic axis. To obtain a continuous control, we propose an elastic range search procedure that automatically identifies an effective interval of steering magnitudes, avoiding both under-steering (no-edit) and over-steering (changing other attributes). Adding the scaled versions of the same vector within this interval yields smooth and continuous edits. Since our method modifies only textual representations, it naturally generalizes across text-conditioned modalities, including image and video generation. To quantify the steering continuity, we introduce a new evaluation metric that measures the uniformity of semantic change across edit strengths. We compare the continuous editing behavior across methods and find that, despite its simplicity and lightweight design, our approach is comparable to training-based alternatives, outperforming other training-free methods.

Finding a Steering Direction

Step 1

Debiased Contrastive Pairs

Generate balanced positive and negative prompts, then keep only the concept-bearing token spans so the direction isolates the intended attribute.

-

A portrait of a frowning person in warm window light.

+

A portrait of a smiling person in warm window light.

-

A close-up shot of a frowning face beside a curtain.

+

A close-up shot of a smiling face beside a curtain.

-

A studio photo of a frowning figure wearing a blue shirt.

+

A studio photo of a smiling figure wearing a blue shirt.

Step 2

Difference-of-Means Direction

Pool the selected token embeddings, average each side, take their difference, and normalize once to obtain the global steering direction ds.

ei+ = pool( E(pi+)[Si+] )
ei- = pool( E(pi-)[Si-] )
s = mean(e+) - mean(e-) , ds = s s2

Difference-of-means over pooled concept-token embeddings.

Animated demonstration of difference-of-means steering and prompt-space application
The pooled negative and positive embeddings define two centroids; their difference becomes the steering axis that is then added back into the prompt representation.

LLM-Assisted Token Selection

The LLM picks only the tokens that should receive the steering vector. This keeps the edit aligned with the concept and avoids drifting the whole prompt. An implicit prompt does not explicitly contain the target attribute to be changed, while an explicit prompt already names that attribute or style in the text.
Local Edit

Add Smile

Implicit Prompt

A portrait of a man in a forest

Steer the main subject noun.

Explicit Prompt

A portrait of a sad man

Steer only the attribute token.

Stylization

Cartoon Style

Implicit Prompt

A lighthouse on a cliff

Steer the main subject noun.

Explicit Prompt

A photorealistic lighthouse on a cliff

Steer the style token.

Selection Rules

Prompts are treated as implicit when the target attribute is absent, and explicit when it is already named.

Edit Type
Implicit Prompt
Explicit Prompt
Local
Steer the main entity.
Steer the attribute token.
Stylization
Steer the main subject noun.
Steer the style token.
Global
Steer the main scene nouns.
Steer only the global edit token.

Elastic Range Search

Given a steering direction, we need a usable range of strengths. We initialize a small set of anchor values, generate outputs for each one, and measure perceptual distances between neighboring edits.

The elastic band procedure then expands the range where gaps are too large, moves interior control points to equalize normalized gaps, and filters strengths that either change too little or distort the image too much relative to the unsteered anchor.

The toy animation below walks through the actual search loop: initialization, gap evaluation, midpoint expansion, left/right moves, and final similarity-based filtering into a usable slider range.

Animated toy run of elastic band search

More Results

We show additional results on Qwen Image Edit and Flux2 models.

The image editing models below as well as the video models in the next section all rely on different text encoders, providing direct evidence that our method generalizes across different backbones.

Flux2

Flux2 cyberpunk result 1 Flux2 cyberpunk result 2 Flux2 cyberpunk result 3 Flux2 cyberpunk result 4 Flux2 cyberpunk result 5 Flux2 cyberpunk result 6

Cyberpunk Scene

Flux2 crowded result 1 Flux2 crowded result 2 Flux2 crowded result 3 Flux2 crowded result 4 Flux2 crowded result 5 Flux2 crowded result 6

Crowdedness

Flux2 ghibli result 1 Flux2 ghibli result 2 Flux2 ghibli result 3 Flux2 ghibli result 4 Flux2 ghibli result 5 Flux2 ghibli result 6

Ghibli Style

Qwen Image Edit

Qwen beard result 1 Qwen beard result 2 Qwen beard result 3 Qwen beard result 4 Qwen beard result 5 Qwen beard result 6

Beard

Qwen rust result 1 Qwen rust result 2 Qwen rust result 3 Qwen rust result 4 Qwen rust result 5

Rust

Qwen night time result 1 Qwen night time result 2 Qwen night time result 3 Qwen night time result 4 Qwen night time result 5 Qwen night time result 6

Night Time

Qwen skinny result 6 Qwen skinny result 5 Qwen skinny result 4 Qwen skinny result 3 Qwen skinny result 2 Qwen skinny result 1

Muscular

Qwen winter result 1 Qwen winter result 5 Qwen winter result 2 Qwen winter result 3 Qwen winter result 4 Qwen winter result 6

Winter

Qwen smile result 1 Qwen smile result 2 Qwen smile result 3 Qwen smile result 4 Qwen smile result 5 Qwen smile result 6

Smile

Results on Wan2.1 and Wan Vace

Because we operate only on the text encoder, the same steering directions can also be applied to video backbones that share the same text encoder.

Wan2.1

Cartoon

Anime

VACE

It is also possible to edit the first frame of a video and then use the VACE pipeline to propagate that edit consistently through the generated video.

Rusty Motorcycle First-Frame Edit

Cartoon First-Frame Edit

BibTeX


        @misc{ekin2026unreasonableeffectivenesstextembedding,
          title={The Unreasonable Effectiveness of Text Embedding Interpolation for Continuous Image Steering}, 
          author={Yigit Ekin and Yossi Gandelsman},
          year={2026},
          eprint={2603.17998},
          archivePrefix={arXiv},
          primaryClass={cs.CV},
          url={https://arxiv.org/abs/2603.17998}, 
      }