/Awesome-ICCV2023-Low-Level-Vision

A Collection of Papers and Codes in ICCV2023/2021 about low level vision

Awesome-ICCV2023-Low-Level-Vision

A Collection of Papers and Codes in ICCV2023 related to Low-Level Vision

[In Construction] If you find some missing papers or typos, feel free to pull issues or requests.

Related collections for low-level vision

Overview

Image Restoration

DiffIR: Efficient Diffusion Model for Image Restoration

Under-Display Camera Image Restoration with Scattering Effect

Multi-weather Image Restoration via Domain Translation

Towards Authentic Face Restoration with Iterative Diffusion Models and Beyond

Improving Lens Flare Removal with General Purpose Pipeline and Multiple Light Sources Recovery

High-Resolution Document Shadow Removal via A Large-Scale Real-World Dataset and A Frequency-Aware Shadow Erasing Net

Physics-Driven Turbulence Image Restoration with Stochastic Refinement

DDS2M: Self-Supervised Denoising Diffusion Spatio-Spectral Model for Hyperspectral Image Restoration

Image Reconstruction

Pixel Adaptive Deep Unfolding Transformer for Hyperspectral Image Reconstruction

Video Restoration

Snow Removal in Video: A New Dataset and A Novel Method

Video Adverse-Weather-Component Suppression Network via Weather Messenger and Adversarial Backpropagation

Fast Full-frame Video Stabilization with Iterative Optimization

Minimum Latency Deep Online Video Stabilization

[Back-to-Overview]

Super Resolution

Image Super Resolution

On the Effectiveness of Spectral Discriminators for Perceptual Quality Improvement

SRFormer: Permuted Self-Attention for Single Image Super-Resolution

Spatially-Adaptive Feature Modulation for Efficient Image Super-Resolution

DLGSANet: Lightweight Dynamic Local and Global Self-Attention Network for Image Super-Resolution

Boosting Single Image Super-Resolution via Partial Channel Shifting

Dual Aggregation Transformer for Image Super-Resolution

Feature Modulation Transformer: Cross-Refinement of Global Representation via High-Frequency Prior for Image Super-Resolution

MetaF2N: Blind Image Super-Resolution by Learning Efficient Model Adaptation from Faces

Lightweight Image Super-Resolution with Superpixel Token Interaction

Iterative Soft Shrinkage Learning for Efficient Image Super-Resolution

Spherical Space Feature Decomposition for Guided Depth Map Super-Resolution

Real-CE: A Benchmark for Chinese-English Scene Text Image Super-resolution

Burst Super Resolution

Towards Real-World Burst Image Super-Resolution: Benchmark and Method

Spatial-Temporal Video Super-Resolution

MoTIF: Learning Motion Trajectories with Local Implicit Neural Functions for Continuous Space-Time Video Super-Resolution

[Back-to-Overview]

Image Rescaling

Downscaled Representation Matters: Improving Image Rescaling with Collaborative Downscaled Images

[Back-to-Overview]

Denoising

Image Denoising

Random Sub-Samples Generation for Self-Supervised Real Image Denoising

Score Priors Guided Deep Variational Inference for Unsupervised Real-World Single Image Denoising

The Devil is in the Upsampling: Architectural Decisions Made Simpler for Denoising with Deep Image Prior

Lighting Every Darkness in Two Pairs: A Calibration-Free Pipeline for RAW Denoising

Unsupervised Image Denoising in Real-World Scenarios via Self-Collaboration Parallel Generative Adversarial Branches

ExposureDiffusion: Learning to Expose for Low-light Image Enhancement

Towards General Low-Light Raw Noise Synthesis and Modeling

Hybrid Spectral Denoising Transformer with Guided Attention

[Back-to-Overview]

Deraining

From Sky to the Ground: A Large-scale Benchmark and Simple Baseline Towards Real Rain Removal

Learning Rain Location Prior for Nighttime Deraining

Sparse Sampling Transformer with Uncertainty-Driven Ranking for Unified Removal of Raindrops and Rain Streaks

[Back-to-Overview]

Dehazing

MB-TaylorFormer: Multi-branch Efficient Transformer Expanded by Taylor Formula for Image Dehazing

Generalizing Event-Based Motion Deblurring in Real-World Scenarios

[Back-to-Overview]

HDR Imaging / Multi-Exposure Image Fusion

MEFLUT: Unsupervised 1D Lookup Tables for Multi-exposure Image Fusion

RawHDR: High Dynamic Range Image Reconstruction from a Single Raw Image

LAN-HDR: Luminance-based Alignment Network for High Dynamic Range Video Reconstruction

[Back-to-Overview]

Frame Interpolation

Video Object Segmentation-aware Video Frame Interpolation

[Back-to-Overview]

Image Enhancement

Iterative Prompt Learning for Unsupervised Backlit Image Enhancement

Low-Light Image Enhancement

ExposureDiffusion: Learning to Expose for Low-light Image Enhancement

Implicit Neural Representation for Cooperative Low-light Image Enhancement

Low-Light Image Enhancement with Illumination-Aware Gamma Correction and Complete Image Modelling Network

Diff-Retinex: Rethinking Low-light Image Enhancement with A Generative Diffusion Model

[Back-to-Overview]

Image Harmonization/Composition

Deep Image Harmonization with Learnable Augmentation

Deep Image Harmonization with Globally Guided Feature Transformation and Relation Distillation

TF-ICON: Diffusion-Based Training-Free Cross-Domain Image Composition

[Back-to-Overview]

Image Completion/Inpainting

Diverse Inpainting and Editing with GAN Inversion

[Back-to-Overview]

Image Stitching

Parallax-Tolerant Unsupervised Deep Image Stitching

[Back-to-Overview]

Image Compression

RFD-ECNet: Extreme Underwater Image Compression with Reference to Feature Dictionary

[Back-to-Overview]

Image Quality Assessment

Delegate Transformer for Image Color Aesthetics Assessment

Test Time Adaptation for Blind Image Quality Assessment

Exploring Video Quality Assessment on User Generated Contents from Aesthetic and Technical Perspectives

[Back-to-Overview]

Style Transfer

AesPA-Net: Aesthetic Pattern-Aware Style Transfer Networks

Two Birds, One Stone: A Unified Framework for Joint Learning of Image and Video Style Transfers

All-to-key Attention for Arbitrary Style Transfer

StyleDiffusion: Controllable Disentangled Style Transfer via Diffusion Models

[Back-to-Overview]

Image Editing

Adaptive Nonlinear Latent Transformation for Conditional Face Editing

Multimodal Garment Designer: Human-Centric Latent Diffusion Models for Fashion Image Editing

MasaCtrl: Tuning-Free Mutual Self-Attention Control for Consistent Image Synthesis and Editing

Not All Steps are Created Equal: Selective Diffusion Distillation for Image Manipulation

HairCLIPv2: Unifying Hair Editing via Proxy Feature Blending

StyleGANEX: StyleGAN-Based Manipulation Beyond Cropped Aligned Faces

Diverse Inpainting and Editing with GAN Inversion

[Back-to-Overview]

Image Generation/Synthesis / Image-to-Image Translation

Text-to-Image / Text Guided / Multi-Modal

MagicFusion: Boosting Text-to-Image Generation Performance by Fusing Diffusion Models

ELITE: Encoding Visual Concepts into Textual Embeddings for Customized Text-to-Image Generation

Better Aligning Text-to-Image Models with Human Preference

Unleashing Text-to-Image Diffusion Models for Visual Perception

Unsupervised Compositional Concepts Discovery with Text-to-Image Generative Models

BoxDiff: Text-to-Image Synthesis with Training-Free Box-Constrained Diffusion

Ablating Concepts in Text-to-Image Diffusion Models

Learning to Generate Semantic Layouts for Higher Text-Image Correspondence in Text-to-Image Synthesis

HumanSD: A Native Skeleton-Guided Diffusion Model for Human Image Generation

Story Visualization by Online Text Augmentation with Context Memory

DiffCloth: Diffusion Based Garment Synthesis and Manipulation via Structural Cross-modal Semantic Alignment

Dense Text-to-Image Generation with Attention Modulation

Image-to-Image / Image Guided

Reinforced Disentanglement for Face Swapping without Skip Connection

BlendFace: Re-designing Identity Encoders for Face-Swapping

General Image-to-Image Translation with One-Shot Image Guidance

GaFET: Learning Geometry-aware Facial Expression Translation from In-The-Wild Images

Scenimefy: Learning to Craft Anime Scene via Semi-Supervised Image-to-Image Translation

Unifying Diffusion Models' Latent Space, with Applications to CycleDiffusion and Guidance

Others for image generation

Conditional 360-degree Image Synthesis for Immersive Indoor Scene Decoration

Masked Diffusion Transformer is a Strong Image Synthesizer

Q-Diffusion: Quantizing Diffusion Models

The Euclidean Space is Evil: Hyperbolic Attribute Editing for Few-shot Image Generation

LFS-GAN: Lifelong Few-Shot Image Generation

FreeDoM: Training-Free Energy-Guided Conditional Diffusion Model

Improving Diversity in Zero-Shot GAN Adaptation with Semantic Variations

Smoothness Similarity Regularization for Few-Shot GAN Adaptation

Video Generation

Bidirectionally Deformable Motion Modulation For Video-based Human Pose Transfer

MODA: Mapping-Once Audio-driven Portrait Animation with Dual Attentions

Text2Video-Zero: Text-to-Image Diffusion Models are Zero-Shot Video Generators

FateZero: Fusing Attentions for Zero-shot Text-based Video Editing

RIGID: Recurrent GAN Inversion and Editing of Real Face Videos

StableVideo: Text-driven Consistency-aware Diffusion Video Editing

StyleInV: A Temporal Style Modulated Inversion Network for Unconditional Video Generation

[Back-to-Overview]

Others [back]

DDColor: Towards Photo-Realistic and Semantic-Aware Image Colorization via Dual Decoders

DDFM: Denoising Diffusion Model for Multi-Modality Image Fusion

Name Your Colour For the Task: Artificially Discover Colour Naming via Colour Quantisation Transformer

Unfolding Framework with Prior of Convolution-Transformer Mixture and Uncertainty Estimation for Video Snapshot Compressive Imaging

Deep Optics for Video Snapshot Compressive Imaging

SimFIR: A Simple Framework for Fisheye Image Rectification with Self-supervised Representation Learning

Single Image Reflection Separation via Component Synergy

Learned Image Reasoning Prior Penetrates Deep Unfolding Network for Panchromatic and Multi-Spectral Image Fusion

Talking Head Generation

Implicit Identity Representation Conditioned Memory Compensation Network for Talking Head video Generation