Thinking in Loops: Scaling Visual ARC with Looped Transformers
Looped (weight-tied) Transformers for Visual ARC—scale reasoning with “thinking time”.
Looped (weight-tied) Transformers for Visual ARC—scale reasoning with “thinking time”.
Difficulty-adaptive RL for diffusion restoration with IQA reward.
Confidence scaling for better autoregressive image generation.
Benchmarking visual reasoning in video generation.
Region-focused RL for human-centric diffusion generation.
Training-free attention scaling for text-to-video semantic fidelity.
Temporal regularization improves video generator quality and stability.
Knowledge distillation + DPO for efficient image generation.
Step-by-step multi-shot video generation with minimal manual intervention.