UT Austin
YouTube LIVE Lab

Perceptual Quality Assessment and Enhancement of Visual Media using Generative Priors.

SHRESHTH SAINI

PhD Student, The University of Texas at Austin

PhD Dissertation Defense

April 2026

What starts here changes the world

Candidate & Committee Members

Shreshth Saini

Candidate

UT-Austin

Prof. Alan C. Bovik (Advisor)

Chair/Member

UT-Austin

Prof. Joydeep Ghosh

Member — UT-Austin

Prof. Diana Marculescu

Member — UT-Austin

Dr. Yan Ye

Member — Alibaba Group US

Dr. Balu Adsumilli

Member — Google Inc.

What starts here changes the world

Dissertation Roadmap

Part I: Perceptual Quality Assessment

Beyond8Bits & BrightRate Review

HDR-UGC datasets & quality models (WACV '24, '26 + ICIP '24)

LGDM: Latent Guidance in Diffusion Models Review

Zero-shot IQA via generative priors (ICML '25)

HDR-Q: MLLM for HDR VQA CVPR '26

First multimodal LLM for HDR quality — HAPO framework

Part II: Generative Enhancement

Rectified-CFG++ NeurIPS '25

On-manifold guidance for rectified flow models

LumaFlux: Inverse Tone Mapping

Physically & perceptually guided SDR→HDR via DiT

Future Directions

Unified perception-generation framework

Unifying theme: Generative models as both evaluators and enhancers of visual quality

PhD DefenseLIVE Lab, UT Austin

Progress Review Recap

Beyond8Bits & LGDM

Establishing HDR-UGC benchmarks and leveraging generative priors for perceptual quality

What starts here changes the world

Review   Beyond8Bits & BrightRate

The largest HDR-UGC video quality dataset & first HDR-aware blind VQA model

44,276HDR-UGC video clips
1.5M+Crowdsourced ratings

BrightRate: Multi-Stream VQA

  • HDR features via expansive non-linearity on extreme luminance
  • UGC features (CONTRIQUE) + CLIP semantic features
  • Temporal difference module for flicker/quality fluctuations
  • SROCC 0.889 vs. prior SOTA 0.853

4-Stage Pipeline

Video Sourcing (Vimeo+Users) → Bitrate Ladder Encoding → AMT Study (35 ratings/video) → SUREAL MOS Aggregation

Beyond8Bits overview

HDR vs SDR gap, Beyond8Bits diversity, and HDR-Q improvements over baselines

PhD DefenseLIVE Lab, UT Austin
What starts here changes the world

Review   LGDM: Latent Guidance for Quality Assessment

Zero-shot NR-IQA by exploiting generative priors — no task-specific training

Core Insight

Pretrained diffusion models encode a perceptual manifold. PMG steers sampling toward perceptually consistent regions — quality assessment without IQA labels.

Key Innovation: Diffusion Hyperfeatures

  • Multi-timestep, multi-layer features from U-Net
  • Theoretically grounded guidance on tangent spaces
  • Works with off-the-shelf SD v1.5 — no fine-tuning
z*t-1 = zt-1 + ζ2 · ∇z Φ(z0|t)

PMG: guide denoising toward perceptually consistent manifold regions

PMG

Perceptual Manifold Guidance — on-manifold quality-aware sampling

LGDM results

LGDM achieves SOTA median SRCC across authentic distortion benchmarks

PhD DefenseLIVE Lab, UT Austin
What starts here changes the world

Dissertation: Five Novel Contributions

Each project introduces new methods, datasets, or frameworks — together they form a complete pipeline from perception to generation for HDR visual media.

Beyond8Bits
HDR Data
LGDM
Gen. Priors
HDR-Q
MLLM + RL
LumaFlux
SDR→HDR
Rect-CFG++
Flow Guidance

Unifying theme: Generative models as both evaluators and enhancers of visual quality

Part I: Perceptual Quality Assessment

HDR-Q: MLLM for HDR VQA

First multimodal LLM for HDR video quality assessment with HDR-Aware Policy Optimization

CVPR 2026

Can AI see and reason about
HDR video quality?

Today's answer: No. Every existing model was built for the 8-bit SDR world.

What starts here changes the world

HDR Is Not "Better SDR" — It's a Different Signal Space

Every modern phone captures 10-bit HDR by default.

YouTube, Instagram, TikTok — billions of daily HDR uploads. HDR10 supports 10-bit depth, BT.2020 wide color gamut, PQ perceptual quantizer. Peak luminance ≥1000 nits vs SDR ~100 nits.

New perceptual phenomena SDR models cannot see:

Highlight clipping · Near-black banding · Color blooming · PQ quantization artifacts · Exposure flicker · Wide-gamut chroma shifts

Core argument:

Standard vision encoders (SigLIP, CLIP) process images in 8-bit sRGB. They are structurally blind to HDR-specific distortions. Not a training gap — a representation gap.

HDR vs SDR

HDR preserves luminance and color detail that SDR collapses

HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

Where Current Methods Break

No existing method combines HDR-aware perception with quality reasoning

Method FamilyHDR InputPercept. Ground.ReasoningCont. MOSInterp.SROCC
Classical (BRISQUE, VMAF)0.41
Deep VQA (FastVQA, DOVER)~0.51
HDR-Specific (HIDRO-VQA)0.85
MLLM-VQA (Q-Insight, DeQA)~0.52
HDR-Q (Ours)0.92

The gap HDR-Q fills:

HDR-aware visual perception + continuous quality prediction + interpretable chain-of-thought reasoning. No prior method has all three.

HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

Three Fundamental Obstacles

Each requires a dedicated solution — together they define the HDR-Q architecture

O1

SDR-pretrained
vision encoders

Blind to 10-bit PQ

O2

Continuous MOS
from token space

Bridging autoregressive & regression

O3

Modality
neglect

GRPO ignores HDR tokens

Evidence: GRPO with HDR input → 0.875 SROCC. SDR-only → 0.891. Adding HDR made it worse.

HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

Beyond8Bits: The Foundation

First large-scale HDR-UGC quality dataset — the training ground for HDR-Q

44,276HDR-UGC video clips
6,861Unique source videos
1.5M+Human quality ratings
~35Ratings per video

Format: 10-bit HEVC, PQ transfer, BT.2020 · 360p–1080p · 0.2–5 Mbps bitrate ladder

Sources: 2,253 crowdsourced (diverse UGC) + 4,608 Vimeo CC (nature, outdoor, low-light)

Quality control: HDR10 display verification · Qualification quiz · Golden set (SROCC 0.85) · SUREAL MOS aggregation · Inter-subject SROCC 0.90

Beyond8Bits

Dataset diversity, HDR vs SDR characteristics, and HDR-Q performance gains

HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

HDR-Q: Architecture Overview

Architecture

Solves O1: HDR-Aware Encoder

SigLIP-2 + dual-domain contrastive learning. Native 10-bit PQ. Dual HDR + SDR pathways.

Solves O3: HAPO Training

Contrastive KL + dual entropy + entropy-weighted advantage. Forces HDR modality attention.

Solves O2: Structured Output

Ovis2.5 (9B) + Rank-4 LoRA. <think> reasoning + <answer> MOS score. Gaussian reward σ=3.

HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

HDR-Aware Vision Encoder

Contrastive HDR/SDR discrimination

Lctr = max(0, δ − D(E(xHDR), E(c)) + D(E(xSDR), E(c)))

Pull HDR close to caption, push SDR away

Full encoder loss

Lenc = LSigmoid + λctr · Lctr

Semantic alignment + HDR discrimination

Design justifications

Why SigLIP-2? Strong semantic priors, multimodal-compatible.
Why contrastive? HDR info = what's in HDR but absent in SDR.
Why 10-bit PQ input? Tone-mapping destroys the signal we need.
Captions by Qwen2.5-VL-72B for quality-aware descriptions.

Encoder training

SigLIP-2 contrastive finetuning with matched HDR-SDR pairs and quality-aware captions

HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

The Modality Neglect Problem

The most surprising finding — and the core motivation for HAPO

HDR-Q (SDR input only)

0.8914

Standard GRPO (HDR+SDR input)

0.8753

Adding HDR made it worse.

Why this happens:

  • GRPO treats all tokens equally — no modality importance signal
  • Autoregressive generation enables text-context shortcuts
  • SDR features are familiar; HDR features are foreign → model ignores the foreign
  • Result: higher input dimensionality, lower information utilization

HAPO fixes this: 0.9206 SROCC

By explicitly rewarding different outputs for HDR vs SDR inputs.

HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

HAPO — Core Mechanism: HDR-SDR Contrastive KL

If the model ignores HDR → identical outputs for HDR & SDR → DKL ≈ 0. We maximize this divergence.

KHDR(θ) = DKLHDRθ ‖ πSDRθ)

Maximized with coefficient γ = 0.5 in the HAPO objective

Formal guarantee:   Iθ*(output; HDR | SDR) ≥ γ · KHDR(θ*) − κ

Mutual information between output and HDR content is lower-bounded. The policy is mathematically guaranteed to use HDR information.

HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

The HAPO Objective

JHAPO(θ) = E[Σ min(ρ · Ã, clip(…) · Ã)] − β DKLHDR ‖ πref) + γ KHDR − Hdual
Ã
Entropy-weighted
advantage (HEW)
λHEW=0.3
β DKL
Reference KL
stability
β=0.02
+γ KHDR
Contrastive KL
HDR grounding
γ=0.5
−Hdual
Dual entropy
anti-collapse
η₁=0.01, η₂=0.05
HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

Training Pipeline & Reward Design

Stage 1: Modality Alignment

Full HAPO with γ=0.5. Curated Beyond8Bits subset with matched HDR-SDR pairs. Goal: teach model to attend to HDR-specific information. Both stages use RL — no SFT stage.

Stage 2: Quality Calibration

Full Beyond8Bits corpus. Score reward as primary signal, reduced γ. Goal: prediction accuracy while preserving HDR grounding from Stage 1.

Why two stages?

Modality alignment and quality calibration are competing objectives. Aggressive contrastive training degrades absolute accuracy; pure quality training enables modality neglect. Sequential prioritization resolves the tension.

Composite Reward

R = wfmt·Rfmt + wsc·Rscore + wself·Rself

Rfmt: Binary — valid <think>/<answer> tags = 1
Rscore: Gaussian — exp(−(ŝ−s*)²/2σ²), σ=3. Smooth, differentiable.
Rself: Majority-vote consistency across K=8 completions.

Config: Ovis2.5 (9B) + Rank-4 LoRA · K=8 · ε=0.1 · 4×H200 · BF16 · T=8 frames · 10-bit PQ native input
HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

Main Results: Beyond8Bits Benchmark

MethodSROCC↑PLCC↑RMSE↓
BRISQUE0.4100.46911.70
CONTRIQUE0.6250.60515.02
CONVIQT0.7990.8108.48
HIDRO-VQA0.8510.8786.09
Q-Insight (best MLLM)0.5170.56220.78
HDR-Q (SDR only)0.8910.8907.42
HDR-Q (Full)0.9210.9125.16
0.921Beyond8Bits
0.908LIVE-HDR
(zero-shot)
0.725SFV+HDR
(zero-shot)
Cross-dataset

Zero-shot generalization without fine-tuning on target datasets

HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

Qualitative: HDR-Q vs Baseline Reasoning

Qualitative

HDR-Q identifies true HDR artifacts (highlight preservation, banding, color fidelity). Ovis2.5 baseline hallucinates non-existent issues.

HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

Ablation: Every Component Is Justified

VariantSROCCRMSETok. H
GRPO baseline0.8110.730.20
+ HDR Encoder0.838.960.24
HAPO w/o Contrastive KL0.867.100.29
HAPO w/o HEW0.886.110.27
HAPO w/o Dual Ent.0.915.820.26
HDR-Q (Full)0.925.150.33

1. Contrastive KL — CRITICAL

0.92 → 0.86 without. Largest single drop. Prevents modality neglect.

2. HEW — Token-level credit assignment

0.92 → 0.88. Directs gradient to quality-relevant tokens.

3. HDR Encoder — Foundation

0.83 → 0.81. Essential for 10-bit PQ representation.

4. Dual Entropy — Stability

0.92 → 0.91. Prevents collapse, maintains exploration.

Token entropy: 0.20 → 0.33. Model becomes more uncertain at quality-critical decisions, not less. This is healthy.

HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

Training Dynamics: HAPO vs GRPO

Entropy

Token entropy during training — GRPO collapses, HAPO maintains healthy exploration

GRPO failure mode

Entropy drops → deterministic → ignores visual modality → text-context shortcuts

HAPO stabilization

Dual entropy maintains H ≈ 0.33 at quality-critical positions. Healthy exploration preserved.

Reasoning efficiency

CoT: 168 → 137 tokens (−18%). More concise, more focused. Boilerplate removed, quality observations retained.

HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

What We Learn & Remaining Challenges

Key insights

  • Modality neglect is the bottleneck, not encoder capacity. The contrastive KL mechanism is necessary and sufficient for HDR grounding.
  • Token-level entropy is a diagnostic signal. Increasing entropy at decision points = better visual grounding. Collapsing entropy = text shortcuts.
  • Two-stage RL outperforms SFT→RL. Both stages use RL, ensuring HDR grounding from initialization.

Remaining challenges

  • Temporal reasoning is limited to T=8 frame sampling. Long-range temporal quality patterns may be missed.
  • Extreme out-of-distribution content (e.g., synthetic HDR, gaming HDR) not well represented in training.
  • Inference cost — MLLM inference is slower than lightweight VQA models. Not yet real-time.
HDR-Q (CVPR '26)LIVE Lab, UT Austin
What starts here changes the world

Key Takeaways

1

First MLLM for HDR video quality. Prior MLLMs: 0.52 SROCC. HDR-Q: 0.92.

2

HAPO: principled solution to modality neglect. Three mechanisms, formal MI guarantee, all ablated.

3

Strong zero-shot generalization. 0.908 LIVE-HDR, 0.725 SFV+HDR — no fine-tuning.

4

Perception-grounded reasoning. CoT references HDR-specific phenomena baseline models cannot articulate.

Next: HDR-Q as reward model for HDR generation & restoration → closing the perception-generation loop.

HDR-Q (CVPR '26)LIVE Lab, UT Austin

Part II: Generative Enhancement

LumaFlux: Inverse Tone Mapping

Physically & Perceptually Guided Diffusion Transformers for SDR→HDR

Can we teach a generative model
the physics of light?

SDR→HDR reconstruction is ill-posed. We need physical priors and perceptual guidance — not just more data.

What starts here changes the world

What Happens When HDR Becomes SDR

Every step is lossy — inverting it is ill-posed

HDR to SDR pipeline
PQ vs Gamma EOTF

Luminance Compression

PQ: 0–10,000 cd/m². Gamma: 100 cd/m². SDR clips 99% of luminance.

CIE gamut BT.709 vs BT.2020

Color Gamut Compression

BT.2020: 75.8% of CIE. BT.709: 35.9%. Over half the color volume discarded.

Forward: xsdr = Γ709(M2020→709(ΓPQ(xhdr)/Lmax)) + ε  —  PQ decodegamut compressgamma encodequantize. Inverting requires generative priors.
LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

What Gets Lost: 10-bit HDR vs 8-bit SDR (Beyond8Bits Dataset)

Split-view from our Beyond8Bits dataset — HDR (upper-left) vs SDR (lower-right) of the same frame

Beyond8Bits HDR vs SDR

Refrigerator scene (top-left): SDR crushes shadow detail inside the fridge — metallic shelving and contents disappear to black. HDR reveals all internal structure.

Autumn lake (top-right): SDR desaturates the vivid fall foliage and loses the cloud detail in the sky. HDR preserves the full color volume and highlight gradations in the water reflection.

Crystal close-up (bottom): SDR clips the specular highlights on the gemstone to flat white. HDR preserves the translucent internal structure and surface reflections.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Real HDR Videos in Yxy Space: What SDR Clips

Computed from real 10-bit PQ BT.2020 videos (Beyond8Bits). Red prism = Rec.709 SDR volume (≤100 nits). Orange dots = pixels outside SDR — lost in tone mapping.

Yxy sunset

Sunset scene: 15% pixels above SDR ceiling. Sun flare + specular highlights on grass reach ~1,300 nits.

Yxy indoor

Indoor/night scene: 27% pixels above SDR ceiling. Bright artificial lights reach ~2,600 nits — massive clipping.

Key insight: 15–27% of HDR pixel information is permanently destroyed by SDR conversion — and the loss is concentrated in the most perceptually important regions: highlights, specular reflections, light sources, and wide-gamut colors. This is what ITM must reconstruct.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Why ITM Is Fundamentally Ill-Posed

Three irreversible losses make exact inversion impossible — any method must hallucinate missing information

1. Many-to-One Mapping

Forward tone mapping is surjective: multiple distinct HDR luminance values (e.g., 500, 2000, 8000 cd/m²) all map to the same clipped SDR code value (255). The inverse has infinite solutions — the mapping is non-invertible.

2. Color Gamut Loss

BT.2020 → BT.709 gamut mapping discards 53% of the color volume. Saturated greens, deep reds, and vivid cyans in the wide gamut are compressed into the sRGB triangle. The original chroma information cannot be recovered from the compressed representation.

3. Quantization Loss

10-bit PQ (1024 levels) → 8-bit gamma (256 levels) destroys fine gradations. In shadows, adjacent PQ code values are 4× further apart in 8-bit, causing visible banding/contouring. In highlights, the entire upper luminance range collapses to a few code values.

Any ITM method must hallucinate plausible HDR detail where SDR has none.

This requires content-aware generative priors — not fixed curves or local CNN features.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Prior ITM Methods: Three Generations of Architectures

From hand-crafted curves to CNNs to diffusion — each generation has fundamental limits

Gen 1: Classical Tone Curves

BT.2446, Reinhard, Huo et al. — Fixed parametric mapping from SDR→HDR using inverse tone curves. Content-blind: the same curve is applied to sunsets, interiors, and neon signs. Cannot recover scene-specific highlights or local contrast.

Gen 2: CNN-Based Learning

HDRTVNet++ (ICCV'21): 3-branch modulation net (global, local, condition). HDCFM (MM'22): Hierarchical feature modulation with dynamic context. HDRTVDM (CVPR'23): Dynamic context transformation. Deep SR-ITM (ICCV'19): Joint super-resolution + ITM.

Limitation: Local receptive fields → cannot model global illumination. Overfit to specific TMOs used in training data.

Gen 3: Diffusion-Based

LEDiff: Latent diffusion conditioned on SDR. PromptIR: Text-guided restoration. FlashVSR: Video super-resolution with diffusion prior.

Limitation: Hue shifts, over-saturated tones, hallucinated details. No physical grounding — the model has no concept of luminance or color space.

Gen 4: LumaFlux (Ours)

Physically & perceptually guided Diffusion Transformer. Frozen 12B Flux backbone + lightweight PGA/PCM modules. First to combine physical priors + perceptual guidance + DiT scale.

HDRTVNet++ architecture

HDRTVNet++ — CNN, local ops only

LEDiff architecture

LEDiff — latent fusion, no physical priors

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Where Prior ITM Methods Break

Classical (Reinhard, BT.2446)

Fixed tone curves. Content-blind. Cannot recover scene-specific highlights.

CNN-based (HDRTVNet++, HDCFM, ICTCPNet)

Limited receptive field. Overfit to fixed tone operators. No global scene understanding.

Diffusion-based (LEDiff, PromptIR)

Color drift & hallucination. Require text prompts or retraining. No physical grounding.

What's missing?

No method uses physical priors (luminance, spectra).
No method uses perceptual priors (semantic understanding).
No method leverages modern DiTs (12B pretrained params).

MethodTypePSNRLimitation
BT.2446ClassicalContent-blind
HDRTVNet++CNN38.36Local receptive field
HDCFMCNN38.42Fixed degradations
LEDiffDiffusion36.52Color drift
PromptIRDiffusion32.14Needs prompts
LumaFluxDiT39.27Physics-guided
LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Prior ITM Methods: Visual Failures

Visual comparison on real HDR dataset — from FastHDRNet (arXiv 2404.04483)

FastHDRNet visual comparison

Desert sunset (top) + evening cityscape (bottom). Ground Truth vs 7 methods: Ada-3DLUT, HDRTVNet, JSI-GAN, CSRNet, HDRNet, HuoPhyEO, FastHDRNet.

Color shift & washout

Ada-3DLUT & HDRTVNet wash out the sun and lose sand gradients. HDRNet introduces severe blue color shift on the cityscape. HuoPhyEO produces unnatural yellow-green cast.

Halo & detail loss

JSI-GAN adds halo artifacts around the sun disc. CSRNet and JSI-GAN lose building detail and city lights in shadow regions. No method recovers both highlights and shadows.

The common pattern

CNN methods lack global context. Analytical methods lack content-awareness. No prior method uses physical priors or perceptual guidance — the exact gap LumaFlux fills.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Background   Flow Matching in 4 Equations

The generation framework behind Flux and LumaFlux

1. Interpolation Path

xt = (1−t)x0 + t·x1

datanoise, straight line

2. Target Velocity

ut = x1x0

Constant — network learns to predict this

3. Training Loss

L = E[‖vθ(xt,t) − (x1x0)‖²]

Simple MSE: predicted vs true velocity

4. Inference: Solve ODE

dx/dt = vθ(x,t)

t=1→0 via Euler. Deterministic, no noise.

Color: data x₀ · noise x₁ · interpolant xₜ · time t · network vθ

Background: Flow MatchingLIVE Lab, UT Austin
What starts here changes the world

Background   Rectified Flow: Straight Paths vs. Noisy Diffusion

Why LumaFlux uses rectified flow (Flux) instead of standard diffusion

Rectified flow vs linear interpolation

From Liu et al. (ICLR '23): (a) Linear interpolation has crossing paths. (b) Rectified flow straightens them via reflow. (c-d) After rectification, paths are straight ≈ optimal transport. Figure from official Rectified Flow repo.

Diffusion (SDE) — Curved, Noisy Paths

Stochastic noise at each step provides implicit error correction. Paths curve and cross. Requires 50+ steps. Models: DDPM, Stable Diffusion 1.5/2.

Rectified Flow (ODE) — Straight, Fast Paths

Deterministic ODE. Paths straightened via reflow ≈ optimal transport. Only 25-40 steps. Consistent outputs. Models: Flux (12B), SD3, SD3.5. LumaFlux builds on Flux.

Why this matters for ITM: Rectified flow gives us deterministic, consistent HDR outputs (no stochastic variation between runs) and fast inference (~8s/frame). The 12B Flux backbone provides rich visual priors from billions of training images — we just need to steer them with physics.

Background: Rectified FlowLIVE Lab, UT Austin
What starts here changes the world

Background   Flux MM-DiT & Why Standard LoRA Isn't Enough

Flux Architecture

12B-param MM-DiT. Visual + text tokens processed jointly via self-attention. Timestep conditioning via adaLN-Zero. Rectified flow ODE, 40 steps.

Standard LoRA

W = W(0) + AB

Rank r=8. Only 0.07% of parameters. Efficient but static.

Why vanilla LoRA fails for ITM

Same adaptation for all inputs, timesteps, layers.
No physics: blind to PQ curves, luminance, spectra.
No perception: can't enforce color constancy.
Result: 33.42 dB PSNR (baseline).

LumaFlux: From LoRA to Luma-MMDiT

Three extensions that make LoRA input-dependent, physically-grounded, and timestep-adaptive:

PGA — Physical gating on value projection

PCM — Perceptual FiLM on hidden states

RQS — Monotone spline tone-field decoder

TLAM — 6 scalars per (timestep, layer)

33.42 → 39.27 dB PSNR (+5.85 dB)

BackgroundLIVE Lab, UT Austin
What starts here changes the world

Architecture Evolution: CNN → MMDiT → Luma-MMDiT

Architecture evolution

Left: CNN stack (local receptive field). Center: Standard Flux MMDiT (global attention but no physical priors). Right: LumaFlux (PGA + PCM + Coupler + RQS).

CNN: Local, no physical priors, no global context

MMDiT: Global attention, 12B priors, but static LoRA, no physics

Luma-MMDiT: + SigLIP + Physical features + spectral gating + RQS

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

LumaFlux: Full Pipeline

Full architecture

PGA
Luminance + gradient + saturation + FFT → gated LoRA on WV

PCM
Frozen SigLIP → cross-attn → FiLM: γ⊙LN(h)+ζ

TLAM
Ψ(t,ℓ) → 6 scalars controlling all modules per step/layer

RQS
K=8 monotone spline in YUV BT.2020. Per-pixel tone curves.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Standard DiT Block → Luma-MMDiT Block

Three targeted modifications — only colored parts are trainable (~17M / 12B)

Block comparison

Left: Standard MMDiT. Right: Luma-MMDiT with PGA on value projection, PCM after norm, Coupler at output.

// Standard MMDiT block

z → LN(z)

Q = zWQ, K = zWK

V = zWV

h = Attn(Q,K,V)

z = z + h + MLP(LN(z))

// + Luma-MMDiT:

V = z(WV0 + Rvt,ℓ) ← PGA

h = γt,ℓ⊙LN(h) + ζt,ℓ ← PCM

z += λ(WpTphys + WcC(Tperc)) ← Coupler

TLAM: Ψ(t, ℓ) → 6 scalars

pga, βpga, αpcm, βpcm, nspec, λ]
Early steps → global tone. Late steps → highlight detail.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

PGA: Physically-Guided Adaptation — Step by Step

Step 1: Extract physical features

Y = [0.2627, 0.6780, 0.0593]T · xlin
Tphys = Conv3×3([Y, log(1+|∇Y|), sat])
g = MLP([μY, σY, p95, p99])

Step 2: Physical & spectral gating

Gphys = diag(σ(Pv[Tphys ‖ g]))
gFFT = softplus(Wr · r)  K=16 bands

Step 3: Modulated LoRA

Rvt,ℓ = (α·AB + β·I) · Gphys · (I + nspec·diag(gFFT))
WV ← WV(0) + Rvt,ℓ

Three multiplicative modulations:

TLAM (α, β) — Timestep-layer scaling. Early = global tone. Late = highlights.

Gphys — Per-head luminance-aware gating. Sigmoid: 0=suppress, 1=pass.

gFFT — Frequency-aware gating. Prevents over-expansion in smooth (low-freq) regions.

Key property: Unlike standard LoRA (static), PGA is input-dependent, timestep-adaptive, and frequency-aware. The same image region gets different adaptation at different diffusion timesteps.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

PCM (Perceptual) & RQS (Decoder)

Perceptual Cross-Modulation

Tperc = φSigLIP(xsdr) (frozen)
t,ℓ, ζt,ℓ] = αpcm · MLP(C(Tperc)) + βpcm
PCM(h) = γ ⊙ LN(h) + ζ

FiLM: scale & shift per dimension after layer norm. Enforces color constancy — e.g., brighten sky highlights without shifting skin tones.

RQS Tone-Field Decoder

Ŷ(α) = [ηk(sk+1)α² + dkskα] / [sk + (sk+1+sk−2dk)α(1−α)]

K=8 knots, per-pixel learned (ξ, η, s).
Monotonic — preserves luminance ordering.
Differentiable & invertible.
Operates in YUV BT.2020 — separates luminance from chroma.

Total Loss

1.0·‖Ylin−Y*‖₁ + 0.5·‖xlin−x*‖₁ + 0.01·Lspline

Spline Smoothness

Lspline = 1/(K−1) Σ(sk+1−sk

Luminance (primary) + color (secondary) + curve smoothness (regularization). All in linear light (PQ EOTF decoded).

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Training: Data & Configuration

~318K paired SDR-HDR images

SourceClipsType
HIDRO-VQA411Professional
CHUG428User-generated
LIVE-TMHDR40Studio + expert SDR

8 TMOs × 3 CRF levels

Reinhard · BT.2446a · BT.2446c+GM · BT.2390+GM · YouTube LogC · OCIOv2 · HC+GM · Expert Graded

Training Configuration

Backbone: Frozen Flux (12B) · Trainable: ~17M params
Optimizer: AdamW · lr=1e-4 · cosine + 5K warmup
Iterations: 200K · batch 16 · 4×H200 · ~48 GPU-hours
Inference: 40 Euler steps · prompt-free · ~8s per 1080p

Evaluation Benchmarks

HDRTV1K
1K pairs, 1080p

HDRTV4K
4K resolution

Luma-Eval
20 unseen · 8 TMOs
New benchmark

Metrics: PSNR (PQ), SSIM, HDR-VDP-3, ΔEITP (ICtCp color difference)

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Results: State-of-the-Art on All Benchmarks

Method1K PSNR↑1K SSIM↑4K PSNR↑VDP3↑ΔEITP
HDRTVNet++38.360.97330.828.758.28
ICTCPNet36.590.92233.128.577.79
HDCFM38.420.97333.258.527.83
LEDiff36.520.87232.255.719.13
PromptIR32.140.95428.489.179.59
LumaFlux39.270.98235.869.836.12
+0.85dB PSNR
HDRTV1K
+2.61dB PSNR
HDRTV4K
9.83HDR-VDP3
(best)
6.12ΔEITP
(best color)

First DiT-based ITM to beat all CNN & diffusion baselines. Strong 4K generalization despite 1080p training. Only ~17M trainable parameters on 12B frozen backbone.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Visual Result 1: Church Interior — Highlight Recovery

Church SDR to HDR Highlight recovered → SDR HDR Zoomed: mosaic detail

SDR Input (left half)

Stained-glass window clips to flat white. Gold ceiling mosaics lose all texture. 8-bit gamma collapses the entire highlight range.

LumaFlux HDR Output (right half)

Window highlights show gradual luminance roll-off. Gold leaf textures fully resolved. Zoomed inset reveals mosaic detail invisible in SDR.

Why this matters

~3 stops of highlight info destroyed by SDR clipping. PGA selectively expands highlights while PCM preserves warm color temperature of the gold mosaic.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Visual Result 2: Shanghai Cityscape — Shadow & Neon Recovery

SDR input (left) vs LumaFlux HDR output (right) — split view with zoomed crop below

Shanghai SDR vs LumaFlux HDR Neon signs restored Tower lights recovered Waterfront detail visible SDR LumaFlux

SDR input (left half)

Pudong waterfront is near-black. Neon signs washed out. Zoomed crop shows boat and tower as indistinguishable dark blobs.

LumaFlux output (right half)

Neon signs vivid with accurate colors. Water reflections reveal full skyline. Zoomed crop shows sharp boat details, tower structure, and signage.

How LumaFlux handles this

PGA spectral gating amplifies shadow signal without artifacts. RQS decoder expands shadows/highlights via per-pixel monotone splines. PCM maps neon colors to correct BT.2020 gamut.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Ablation: Every Component Is Justified

Progressive addition of components on Luma-Eval benchmark (100K iterations)

ConfigurationPSNR↑ΔEITPVDP3↑ΔPSNR
Flux + LoRA33.428.587.82
+ PGA (no spectral)34.947.628.18+1.52
+ spectral gating35.187.318.29+0.24
+ PCM35.896.788.46+0.71
+ RQS (linear)35.726.858.41−0.17
+ RQS (monotone)35.986.098.61+0.26

Total: +2.56 dB over vanilla LoRA baseline

ΔPSNR contribution breakdown

PGA (phys.)
+1.52
Spectral
+0.24
PCM
+0.71
RQS
+0.09

Key findings

PGA is the largest contributor (+1.52 dB) — physical luminance features are critical. PCM adds +0.71 dB — perceptual semantics complement physics. RQS monotone > linear — monotonicity constraint prevents tone reversal and improves ΔEITP by 0.76.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

User Study: Expert Perceptual Evaluation

10 experts, 10 HDR clips, 5-point MOS on calibrated HDR display (LG OLED, 1000 nits)

Brightness 3.8 3.1 3.5 2.8 Color 4.5 4.0 3.6 2.9 Overall 4.2 4.0 3.5 2.7 LumaFlux LEDiff HDRTVNet++ BT.2446c

LumaFlux wins on all three criteria

Color fidelity: 4.5/5 — experts noted accurate neon reproduction, no hue shifts, stable skin tones. Overall: 4.2/5 — highest across brightness, color, and overall quality.

LEDiff is competitive on overall (4.0)

But loses on brightness (3.1 vs 3.8) — diffusion hallucination produces plausible-looking but physically inaccurate luminance. LumaFlux's PGA prevents this.

Classical BT.2446c trails on everything

Content-blind tone curves cannot compete with learned, physically-guided reconstruction. Overall: 2.7/5.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

What We Learn & Remaining Challenges

Key insights

  • Physical conditioning is the largest contributor (+1.52 dB from PGA alone). Luminance-aware gating is more valuable than semantic priors for ITM.
  • Spectral gating prevents over-expansion in smooth regions. Frequency analysis is a natural complement to spatial features.
  • Monotone RQS outperforms linear expansion. Enforcing physical constraints (monotonicity) improves both metrics and perceptual quality.
  • ~17M parameters suffice to adapt a 12B backbone. The pretrained visual priors in Flux are extremely powerful — the key is steering them with physics.

Remaining challenges

  • No temporal consistency — frame-by-frame processing. Video ITM needs temporal attention.
  • VAE bottleneck — Flux VAE is trained for SDR. Fine-grained HDR detail may be lost in latent space.
  • Inference speed — 40 ODE steps at ~8s/frame. Not yet real-time. Distillation is a clear next step.
LumaFlux (NeurIPS '26)LIVE Lab, UT Austin
What starts here changes the world

Key Takeaways

1

First physically-guided DiT for inverse tone mapping. Luminance, gradients, spectra injected directly into attention via gated LoRA.

2

~17M trainable on 12B frozen. Parameter-efficient adaptation that preserves Flux's visual priors while adding physical grounding.

3

SOTA on all benchmarks, all metrics. +0.85 dB HDRTV1K, +2.61 dB HDRTV4K. 91% human preference over classical methods.

4

RQS decoder ensures physically valid tone expansion. Monotone, differentiable, per-pixel — eliminates banding while preserving highlights.

Next: Temporal LumaFlux for video · Model distillation for real-time · HDR-Q as perceptual reward model for LumaFlux → closing the perception-generation loop.

LumaFlux (NeurIPS '26)LIVE Lab, UT Austin

Part II: Generative Enhancement

Rectified-CFG++: Geometry-Aware Flow Guidance

On-manifold predictor-corrector guidance for rectified flow models

NeurIPS 2025

Why does guidance break on the
best image generators?

Flux, SD3, SD3.5 use rectified flows — deterministic ODE.
Standard CFG pushes trajectories permanently off-manifold.

What starts here changes the world

Why Do We Need Guidance At All?

Without guidance, flow models generate plausible but generic images that loosely match the prompt

The problem with unconditional generation

Models trained with conditioning dropout learn both p(x|y) and p(x). Without guidance, outputs come from a mixture of conditional and unconditional — resulting in low prompt adherence, muted details, and bland compositions.

What guidance should do

Amplify the conditional signal — push generation toward what the prompt describes. Sharper details, more vivid colors, better text-image alignment. This is what CFG achieves in diffusion models (DDPM, SDXL).

What actually happens on flow models

CFG was designed for stochastic SDE samplers. Flow models use deterministic ODE. The extrapolation that works in SDEs becomes catastrophic in ODEs — artifacts, color blow-out, structural distortion.

Flux-dev: Same prompt, same seed

No guidance
CFG ω=3.5
Ours λ=3.5

No guidance: dark, muddy, poor prompt adherence. CFG: oversaturated, unnatural. Ours: clean, natural, faithful.

No guidance
CFG ω=3.5
Ours λ=3

SD3.5: Same pattern — CFG oversaturates, ours stays natural.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

CFG Artifacts on Flow Models: A Gallery of Failures

All images generated with standard CFG on flow models. These are not cherry-picked — this is what CFG does to every prompt.

Flux-dev with CFG ω=3 (from paper Figure 6)

Plastic / cartoon
Oversaturated
Flat / clipart
Blown highlights
Washed out

SD3 with CFG ω=3–3.5

Oversaturated
Text garbled
Misspelled
Color blowout
Color distortion

Common failure modes: (1) Oversaturation — colors beyond natural range. (2) Cartoonification — photorealistic becomes plastic. (3) Text corruption — letters distort. (4) Detail blowout — textures replaced by flat patches.

Not rare failures. CFG artifacts appear on every prompt across Flux, SD3, and SD3.5. The problem is fundamental: CFG extrapolation is incompatible with deterministic ODE.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Root Cause: Why CFG Breaks on Flows but Works on Diffusion

The same extrapolation trick, but fundamentally different samplers

Obviously broken: CFG on SD3 & Flux

SD3: "CyberCore Cafe"
SD3: "Entropy and the Soul"
Flux: plastic/cartoon

Garbled text ("Cyberre Cidie"), misspellings ("Entφy amd"), plastic cartoonification — all from standard CFG.

SDE Sampler (Diffusion) vs ODE Sampler (Flows)

SDE (DDPM): +σε +σε +σε Noise corrects! ODE (Flux): Manifold drift No return!

Diffusion SDEs: built-in safety net

Each step adds Gaussian noise — the renoising step. This noise accidentally provides error correction: even when CFG pushes off-manifold, the stochastic noise partially pulls the sample back toward the learned distribution.

dx = [f(x,t) − g(t)²·∇log pt(x)]dt + g(t)·dw̄   ← stochastic noise term

Discrete (DDPM): xt-1 = μθ(xt, t) + σt·z,   z ~ N(0, I). The σt·z term adds fresh noise at every step — pulling samples back toward the learned distribution.

Flow ODEs: no correction mechanism

Rectified flows use a purely deterministic ODE. Once CFG pushes the trajectory off the manifold at step t, there is no mechanism to return. The off-manifold point becomes the starting point for step t+1.

dx = vθ(x, t)·dt          ← purely deterministic, no noise term!

Compounding error

At each step, CFG evaluates guidance at an off-manifold point — making the velocity direction even more wrong. Errors accumulate monotonically across all N sampling steps. The more steps, the worse it gets.

This is why we need a fundamentally different approach — not a fix for CFG, but a new guidance paradigm designed specifically for deterministic ODE samplers.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

What Would an Ideal Guidance Method Look Like?

Desiderata for guidance on rectified flow models

D1: Stay on the manifold

The guided trajectory must remain in a bounded neighbourhood of the learned transport manifold Mt. No off-manifold drift, no error accumulation.

D2: Adaptive — strong early, gentle late

Early steps determine global structure (layout, color palette) — guidance should be strong. Late steps resolve fine details (text, textures) — guidance should vanish to prevent corruption.

D3: Provable guarantees

Not just empirically better — we want formal bounds on manifold distance and distributional deviation. No prior guidance method provides this.

D4: Drop-in replacement

Works with any pretrained flow model — Flux, SD3, SD3.5, Lumina — without retraining. Just swap the sampling algorithm.

Do existing methods satisfy these?

MethodD1D2D3D4
CFG
CFG-Zero*
APG~
Rect-CFG++ (Ours)

Our key insight: Don't extrapolate the guidance direction from the current point. Instead, predict a midpoint on the manifold, evaluate guidance there, and apply it as a bounded additive correction to the conditional velocity.

Three ideas → three guarantees:

Predictor half-step (on-manifold evaluation) + Adaptive α(t) (vanishing guidance) + Bounded correction (interpolation not extrapolation)

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

The Geometric Insight: Why CFG Fails and How We Fix It

Full geometric diagram

Three panels: (Left) Ideal flow on manifold. (Center) CFG extrapolates off-manifold. (Right) Rect-CFG++ stays on-manifold via predictor-corrector. Let's walk through each.

Why does CFG work in diffusion but not in flows?

In diffusion models (SDE samplers like DDPM), each step adds stochastic noise — the renoising step. This noise acts as implicit error correction: even if CFG pushes the trajectory off-manifold, the added noise partially brings it back. It's a built-in safety net.

In rectified flow models (ODE solvers like Flux), sampling is purely deterministic — there is no renoising step. Once CFG pushes the trajectory off the manifold, there is no mechanism to return. The error at step t becomes the starting point for step t-1, and the next CFG extrapolation makes it worse. Errors accumulate monotonically across all sampling steps.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Geometric Insight (1/3): The Ideal Flow on the Manifold

All three paths comparison
Method legend Flow legend

What you're seeing

Three manifolds: Mt (current time, blue/gold band), Mt-1 (next step), M0 (data, orange curve at bottom). The red dots are the ideal positions at each timestep.

The ideal trajectory

The ideal sample flows along the manifold surface from Mt → Mt-1 → M0. At each step, the velocity lies in the tangent space of the manifold — never leaving it.

Three flows shown

Blue = Flux conditional (vc), Red = CFG (extrapolated), Green = Rectified-CFG++ (ours). Notice how green and blue stay near the manifold while red diverges.

Key question

How do we guide the generation (improve text adherence) while staying on the manifold? Let's see what goes wrong first...

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Geometric Insight (2/3): CFG Extrapolates Off-Manifold

CFG off-manifold
Method legend

What CFG does

Starting from the black dot on Mt, CFG computes the guided velocity v̂ω at the current point. With ω>1, this velocity extrapolates past the conditional velocity vc.

ω = vu + ω(vc − vu)    ω > 1 → overshoots!

The tangent space violation

The guided velocity v̂ω does not belong to the tangent space TxtMt. The resulting Euler step lands the sample off Mt-1. In a deterministic ODE, there is no mechanism to return.

The compounding effect

At the next step, CFG evaluates guidance at the off-manifold point — making the direction even worse. Errors accumulate at every step. By the final image: oversaturation, color blow-out, structural distortion.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Geometric Insight (3/3): Our Predictor-Corrector Stays On-Manifold

Our predictor-corrector on-manifold
Method legend

Step 1: Predict midpoint with vc only

Take a half-step using only the conditional velocity: x̃mid = xt + (Δt/2)·vct. Since vc lies in the tangent space, x̃mid lands on or near Mt-Δt/2.

Step 2: Evaluate guidance at the predicted midpoint

Compute vcmid and vumid at x̃mid. The guidance direction Δvmid = vcmid − vumid is evaluated on the manifold, reflecting the geometry of the target manifold Mt-Δt.

Step 3: Apply bounded interpolative correction

v̂ = vct + α(t)(vcmid − vumid)

α(t) = λmax(1-t)γ. Base velocity is always vc (on manifold). Correction is additive and bounded. α(t)→0 near data → fine details preserved.

Result: trajectory stays in bounded tubular neighbourhood of Mt

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

The Core Distinction: Extrapolation vs. Interpolation

This single difference explains why CFG fails and Rect-CFG++ works

CFG: Extrapolation (ω > 1)

Mt xt vu vc ω PAST vc ! Off-manifold zone (ω−1)Δv = extrapolation 0 (vu) 1 (vc) ω (>1)

CFG moves PAST the conditional direction.
ω=3 means 3× the distance from vu to vc. Overshoots into unknown territory.

Ours: Interpolation (0 ≤ α ≤ 1)

Mt xt Δt/2 · vc mid vc (base, on manifold) +α·Δvmid t-1 Still on manifold! α(t)·Δv = small, bounded vc (base) α ∈ [0, 1] vc + Δv ← never goes past this

Ours stays BETWEEN vc and vc+Δv.
Base is always vc (on manifold). Correction α(t)·Δvmid is bounded and decays to 0 near data.

Mathematical contrast:   CFG: (1−ω)vu + ωvc  with ω∈[3,9] → extrapolation far past vc   |   Ours: vc + α(t)Δvmid  with α∈[0,1] → bounded additive correction to vc
Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Algorithm: CFG vs. Rectified-CFG++ (Side-by-Side)

Color highlights show exactly what changes — green = our additions

Standard CFG Sampling

for t = T down to 1:

  vc = vθ(xt, t, y)

  vu = vθ(xt, t, ∅)

  v̂ = vu + ω(vc − vu) ← extrapolation!

  xt-1 = xt + Δt · v̂

2 NFE/step. ω>1 pushes off manifold. No correction.

Rectified-CFG++ (Ours)

for t = T down to 1:

  vc = vθ(xt, t, y)

  mid = xt + (Δt/2)·vc ← predictor

  vcmid = vθ(x̃mid, t-Δt/2, y)

  vumid = vθ(x̃mid, t-Δt/2, ∅)

  v̂ = vc + α(t)(vcmid − vumid) ← interpolation!

  xt-1 = xt + Δt · v̂

3 NFE/step. α(t)=λ(1-t)γ decays → on-manifold. Bounded.

Key differences: (1) Base velocity is always vc, not extrapolated. (2) Guidance evaluated at predicted midpoint on manifold, not current point. (3) Adaptive α(t) → 0 near data.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Guidance Methods Compared: Equations & Properties

Four approaches to the same problem — only ours avoids extrapolation entirely

Standard CFG (Ho & Salimans, 2022)

v̂ = (1 − ω)·vu(xt) + ω·vc(xt)

ω > 1 → extrapolation past vc. Designed for SDEs with renoising. On ODEs: off-manifold drift, oversaturation, error accumulation. Constant guidance at all timesteps.

CFG-Zero* (Wang et al., 2024)

v̂ = (1 − ω)·st*·vu(xt) + ω·vc(xt)

Adds an optimal rescaling factor st* to align unconditional/conditional norms. Reduces initial drift but still extrapolates (ω > 1). No late-step adaptation — fine details still corrupted.

APG (Sadat et al., 2024)

v̂ = vc(xt) + Δvt(η,r,β)

Projects guidance direction onto perpendicular subspace to reduce parallel component. Partial mitigation but guidance still evaluated at current point (possibly off-manifold). No decay schedule, no formal bounds.

Rectified-CFG++ (Ours)

v̂ = vc(xt) + α(t)·(vcmid − vumid)

Three key differences: (1) Base = vc always (on-manifold). (2) Guidance at predicted midpointmid (on-manifold). (3) α(t) = λ(1-t)γ → 0 near data. Bounded, interpolative, provable.

PropertyCFGCFG-Zero*APGOurs
Guidance typeExtrapolationExtrapolationProjectionInterpolation
Evaluated atCurrent xtCurrent xtCurrent xtPredicted x̃mid
Temporal scheduleNone (constant)NoneNoneα(t) → 0
Formal guarantees✓ (3 props)
NFE per step2223
Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Theoretical Guarantees — First Provable Bounds for Flow Guidance

Four mild assumptions → three rigorous guarantees. No prior guidance method has these.

Assumptions

(A1) vθ(x,t,y) and vθ(x,t,∅) are L-Lipschitz in x

(A2) Guidance direction bounded: ‖Δvtθ(x)‖ ≤ B

(A3) Schedule α(t) is bounded and integrable

(A4) Conditional velocity bounded: ‖vct‖ ≤ Vmax

All standard regularity conditions. Lipschitz and boundedness are satisfied by any well-trained neural network. No exotic assumptions needed.

Lemma 1: Guidance Stability

‖Δvθt-Δt/2 − Δvθt(xt)‖ ≤ L · Vmax · Δt

Guidance direction at the predicted midpoint differs from guidance at current point by only O(Δt). The predictor doesn't introduce large errors.

Proposition 1: Bounded Single-Step Perturbation

‖x̂t-1 − x̃t-1‖ ≤ α(t) · B · Δt

Deviation from the pure conditional path is controlled by α(t) at each step. Since α(t) → 0 as t → 0, perturbations vanish near the data manifold.

Proposition 2: Distributional Deviation

DKL(p̂₀‖p₀) ≤ Cs·(B + LVmaxΔt)·∫α(τ)dτ

Total KL divergence controlled by integrated guidance ∫α(τ)dτ — the single quantity that determines output quality.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Why These Guarantees Matter: CFG vs. Ours

Lemma 2: Manifold-Faithful Corrector

dist(x̂, Mt-Δt) ≤ C·ε·Δt

Distance to manifold bounded by training error ε only. The corrector displacement is tangent to Mt-Δt/2 — guidance doesn't push off-manifold.

Integrated Guidance: The Key Quantity

CFG: ∫(ω−1)dt = ω−1 ≈ 2–8 × 0.5 Ours: ∫α(τ)dτ = λ/(γ+1) ≈ 0.5 4–16× lower KL divergence

Head-to-Head: Every Property

PropertyCFGOurs
Per-step perturbation(ω−1)·B·Δt
constant, never vanishes
α(t)·B·Δt
→ 0 near data
Manifold distanceUnbounded
errors accumulate
O(ε·Δt)
training error only
Integrated guidanceω−1 ≈ 2–8λ/(γ+1) ≈ 0.5
KL(p̂₀ ‖ p₀)O(ω−1)O(0.5)
Late-step behaviorSame ω everywhere
→ destroys fine details
α(t) → 0
→ preserves fine details
Guidance evaluationAt current point
(possibly off-manifold)
At predicted midpoint
(on manifold)

Bottom line: CFG has no convergence guarantee for flow models. Ours proves that the output distribution deviates from the true conditional distribution by a controllable, bounded amount — proportional to ∫α(τ)dτ which we set to ≈0.5.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Intermediate Denoising: Watching Artifacts Grow vs. Stay Clean

Each strip shows 7 denoising steps from noise (left) to final image (right)

Standard CFG (ω=2.5) — artifacts compound

Rectified-CFG++ (λ=0.5) — clean throughout

CFG: Off-manifold drift begins at early steps (step 2-3). By mid-trajectory, color artifacts are baked in. Late steps cannot correct — each step amplifies the error. The final image has oversaturated colors and unnatural contrast.

Ours: Predictor step keeps each intermediate on Mt. α(t) schedule applies strong guidance early (global structure) and vanishes late (fine details). Result: every intermediate looks natural. No error accumulation.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

2D Toy: CFG Drifts Off-Support, Ours Stays On-Manifold

200 trajectories on 2D mixed Gaussian. Top: CFG (blue) drifts off support. Bottom: Ours (green) smooth on-manifold.

CFG (top)

Trajectories initially overshoot, leaving the learned transport manifold. Sharp late-stage corrections pull samples back — but damage is done. Final distribution is noisy and off-center.

Rectified-CFG++ (bottom)

Smooth convergence throughout. Predictor anchors each step on the flow field. Corrector applies bounded guidance. Samples arrive at target with tight concentration — no drift, no sharp corrections.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Flux: Baseline vs. CFG vs. Rectified-CFG++ (3 Prompts)

3x3 comparison

Top: Flux baseline (no guidance). Middle: CFG — oversaturated, cartoonish, structural distortion. Bottom: Rect-CFG++ — sharp details, faithful colors, correct structure.

Cactus: CFG turns photorealistic cactus into cartoon with emoji-like face. Ours preserves desert realism.

Dog + moon gate: CFG loses the stone arch, replaces with painted mountains. Ours preserves the gate structure and moon.

Parrot: CFG turns the bird into a psychedelic abstraction. Ours preserves natural plumage and forest setting.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Flux: Baseline vs. CFG vs. Rectified-CFG++ (Same Prompt, Same Seed)

Same model (Flux-dev), same prompt, same seed — three guidance strategies. Left: no guidance (baseline). Center: CFG (ω=3.5). Right: Rectified-CFG++ (λ=0.5). Ours: sharper details, faithful colors, no rainbow artifacts.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Diverse Prompt Gallery: Rectified-CFG++ Handles Everything

All images generated with Rectified-CFG++ on Flux-dev. No cherry-picking — random prompts from MS-COCO and Pick-a-Pic.

Diverse gallery

Photorealistic scenes

Landscapes, portraits, food, architecture — natural lighting, correct shadows, no oversaturation. The on-manifold property preserves the photorealistic distribution Flux was trained on.

Artistic & stylized

Oil paintings, digital art, fantasy scenes — style transfer works correctly because α(t) allows strong guidance early (global style) while preserving fine brush strokes late.

Text & compositional

Signage, book titles, multi-object scenes — the hardest category for any guidance method. α(t)→0 at late steps preserves pixel-precise text rendering and spatial relationships.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

SD3.5: 4-Way Guidance Comparison (Same Prompt, Same Seed)

CFG ω=3.5
CFG-Zero*
APG
Ours λ=6
CFG
CFG
Ours
Ours
Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Flux-dev: CFG vs. Ours (8 Comparisons)

CFG
Ours
CFG
Ours
CFG
Ours
CFG
Ours
Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

More Visual Comparisons: SD3 & Lumina-Next

SD3 CFG
SD3 Ours
SD3 CFG
SD3 Ours
Lumina CFG
Lumina Ours
Lumina CFG
Lumina Ours

Drop-in replacement across all rectified flow models — no retraining, no arch changes.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Text Rendering: The Stop Sign Test

Text-heavy prompts are especially sensitive to off-manifold drift at late denoising steps

CFG

Rectified-CFG++

Prompt: "A stop sign with 'ALL WAY' written below it."

CFG: "STOP" text distorts — letter shapes warp, "ALL WAY" becomes unreadable. Off-manifold drift corrupts high-frequency text strokes in the final denoising steps.

Ours: Crisp, photorealistic stop sign. α(t)→0 near t=0 ensures pure conditional flow during fine detail resolution — text strokes remain pixel-precise.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Text Rendering: Neon Street Sign

CFG

Rectified-CFG++

Prompt: "A neon street sign that says 'CyberCore Cafe', glowing in magenta and blue."

CFG garbles the neon letters — "Cyberre Cidie Cafe" instead of "CyberCore Cafe". Glow halos and letter boundaries bleed together. Ours renders each letter distinctly with clean neon glow separation.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Text Rendering: Headlines & Inscriptions

CFG

Ours

CFG

Ours

Prompt: "A crow detective reading a paper titled 'Feathered Conspiracies', headline in bold noir font."

Prompt: "A magical sword embedded in stone, with the name 'SOLARFANG' etched along its blade in glowing runes."

CFG: "Feathered Conspiracies" → "Feathhrad Conspiracies". "SOLARFANG" → garbled runes. Ours: Both headlines render correctly with proper typography — even on complex surfaces like aged paper and glowing stone.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Text Rendering: Presentations & Artistic Text

CFG

Ours

CFG

Ours

Prompt: "A fox giving a TED talk, slide behind reads 'Entropy and the Soul'."

Prompt: "A burning scroll with the title 'Soulbound by Nightfall' in ornate calligraphy."

CFG: "Entropy and the Soul" → "Entφy amd the Soul". Title on scroll barely readable. Ours: Clean title text on both the presentation screen and the burning scroll — fine details preserved by zero guidance at final steps.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Text Rendering: Carved & Printed Text

CFG

Ours

CFG

Ours

Prompt: "A desert gate carved in sandstone reading 'CITY OF WHISPERS'."

Prompt: "A Japanese fish stamp print with 'WAVES OF JUSTICE' in red ink."

Pattern: Across all text types — neon, carved stone, ink stamps, calligraphy, digital displays — Rectified-CFG++ preserves legibility. The adaptive schedule α(t)→0 is key: no guidance perturbation during fine-detail resolution.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Text Legibility: Full Gallery (SD3.5)

Each column: CFG left → Rectified-CFG++ right

CFG
Ours
CFG
Ours
CFG
Ours
CFG
Ours
CFG
Ours
CFG
Ours
CFG
Ours
CFG
Ours

Consistent text legibility across all prompts — the adaptive schedule α(t) is the key mechanism.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Quantitative: MS-COCO 10K — Consistent Improvements

ModelMethodFID↓CLIP↑ImgRwd↑Pick↑HPS↑
LuminaCFG26.930.2610.54721.030.253
LuminaOurs22.490.2680.62121.190.259
SD3CFG26.330.2730.68421.480.264
SD3Ours24.680.2780.75321.620.269
FluxCFG37.860.2850.89222.040.279
FluxOurs32.230.2890.96122.180.283
Guidance (SD3.5)FIDImgRwdCLIPHPSv2
CFG67.711.0530.3520.294
CFG-Zero*68.390.9950.3460.288
APG67.231.0750.3510.294
Rect-CFG++67.151.0850.3510.296

Best or tied on every metric, every model.

-5.6FID on Flux
-4.4FID on Lumina
4Models tested
0Extra compute
Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

User Study: 30 Experts × 32 Prompts × 4 Models = 15,360 Responses

User study per model

4-way forced choice: CFG (dark blue), APG (medium blue), CFG-Zero* (light blue), Rectified-CFG++ (green). 4 criteria × 4 models. Green bar is tallest in every single comparison.

Overall: 34-36% preference share across all models (random chance = 25%). p < 0.001 vs second-best (APG).

Text legibility: Highest improvement dimension — 28-32% vs 22-25% for alternatives. α(t) schedule is critical.

Protocol: Fleiss' κ = 0.61 (substantial agreement). Images randomized. Experts with CV/generative AI knowledge.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Ablation 1: Guidance Scale — CFG Crashes, Ours Stays Stable

Guidance scale ablation

X-axis: guidance scale (λ for ours | ω for CFG). Y-axis: FID↓, CLIP↑, ImageReward↑, Aesthetic↑. Blue = CFG, Green = Ours.

CFG (blue curves)

FID explodes: 77 → 148 at ω=10. ImageReward collapses: 1.0 → -1.5. CLIP drops 30%. Aesthetic drops 35%. Catastrophic failure at high guidance.

Rectified-CFG++ (green curves)

FID: 75 → 75 (flat!). ImageReward: 1.0 → 0.95. CLIP stable. Aesthetic stable. Graceful degradation — never crashes.

Why this matters

CFG requires careful per-model tuning of ω to avoid collapse. Rect-CFG++ is robust across a 50× range of λ — no hyperparameter sensitivity. One setting works everywhere.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Ablation 2: Sampling Steps — Better Quality with Fewer Steps

Steps ablation

X-axis: sampling steps. Blue = CFG, Green = Ours. Our method converges faster across all metrics.

Ours at 5 steps vs CFG at 5 steps

FID: 71 vs 178 (2.5× better). ImageReward: 1.0 vs -1.5. At ultra-low NFE, CFG produces garbage; ours produces usable images.

Ours at 15 steps ≈ CFG at 28 steps

Same FID, same CLIP, same ImageReward — but ~2× faster inference. This is because our on-manifold trajectories converge more efficiently.

Component ablation (SD3.5)

ConfigFIDCLIPHPSv2
Unconditional only91.120.1440.187
w/o Predictor73.700.3410.297
w/o Corrector74.650.3410.298
Full72.970.3450.300

Both predictor and corrector essential.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Key Takeaways

1

First on-manifold guidance for rectified flow models. Predictor-corrector with adaptive decay. Provably bounded manifold distance.

2

4-16× tighter distributional bounds than CFG. Three propositions with formal proofs. Integrated guidance λ/(γ+1) vs ω-1.

3

Universal drop-in, zero overhead. Flux, SD3, SD3.5, Lumina — all improve. No retraining. 15 steps ≈ CFG at 28.

4

43.5% preference, 15K expert responses. Best on detail, color, text, overall across all models. p < 0.001.

Next: Rect-CFG++ for video generation (Sora-class) · Distilled flow models · Integration with LumaFlux for guided HDR generation.

Rectified-CFG++ (NeurIPS '25)LIVE Lab, UT Austin
What starts here changes the world

Summary of Contributions

1

HDR-Q

First MLLM for HDR VQA

HAPO: 3 novel RL mechanisms

Formal MI guarantee

0.921 SROCC

+8.2% over prior SOTA

2

Rect-CFG++

First on-manifold guidance for flows

Predictor-corrector + adaptive decay

Provable bounded deviation

43.5% preference

Drop-in, zero overhead

3

LumaFlux

First physics-guided DiT for ITM

PGA + PCM + RQS decoder

17M trainable / 12B frozen

+0.85 dB PSNR

SOTA on all 3 benchmarks

Foundation: Beyond8Bits (44K HDR videos, 1.5M+ ratings) · LGDM (zero-shot IQA, ICML '25) · HIDRO-VQA · BrightRate

PhD DefenseLIVE Lab, UT Austin
What starts here changes the world

Connecting the Contributions

Understanding perception enables guiding generation, which enhances perception.

Perceive

Beyond8Bits · LGDM
HDR-Q (MLLM + HAPO)

Guide

Rect-CFG++
On-manifold predictor-corrector

Enhance

LumaFlux
PGA + PCM + RQS

Generative priors are the common thread — pretrained models encode rich perceptual knowledge for both assessment and enhancement.

PhD DefenseLIVE Lab, UT Austin

Publications

  1. 1. TABES: Trajectory-Aware Backward-on-Entropy Steering for Masked Diffusion Model [Under Review]
  2. 2. LumaFlux: Lifting 8-Bit Worlds to HDR Reality with Physically-Guided Diffusion Transformers [Under Review]
  3. 3. Seeing Beyond 8bits: Subjective and Objective Quality Assessment of HDR-UGC Videos [CVPR 2026]
  4. 4. BrightRate: Quality Assessment for User-Generated HDR Videos [WACV 2026] (Oral / Best Paper Award)
  5. 5. Rectified CFG++ for Flow Based Models [NeurIPS 2025]
  6. 6. LGDM: Latent Guidance in Diffusion Models for Perceptual Evaluations [ICML 2025]
  7. 7. CHUG: Crowdsourced User-Generated HDR Video Quality Dataset [IEEE ICIP 2025]
  8. 8. HIDRO-VQA: Deep Contrastive Representation Learning for HDR Video Quality Assessment [WACV 2024]
Mr Bean

Questions?

Thank You!

Perceptual Quality Assessment and Enhancement
of Visual Media using Generative Priors

Shreshth Saini

PhD Candidate

✉   shreshth@utexas.edu

🌎   shreshthsaini.github.io

►   Slides: shreshthsaini.github.io/slides/defense.html

Laboratory for Image & Video Engineering (LIVE)

The University of Texas at Austin

Advised by Prof. Alan C. Bovik