dreamplayer-zhang / Awesome-CVPR2023-Low-Level-Vision

A Collection of Papers and Codes in CVPR2023/2022 about low level vision

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Awesome-CVPR2023-Low-Level-Vision

A Collection of Papers and Codes in CVPR2023 related to Low-Level Vision

[Completed] If you find some missing papers or typos, feel free to pull issues or requests.

Related collections for low-level vision

Overview

Image Restoration

Efficient and Explicit Modelling of Image Hierarchies for Image Restoration

Comprehensive and Delicate: An Efficient Transformer for Image Restoration

Learning Distortion Invariant Representation for Image Restoration from A Causality Perspective

Generative Diffusion Prior for Unified Image Restoration and Enhancement

DR2: Diffusion-based Robust Degradation Remover for Blind Face Restoration

Bitstream-Corrupted JPEG Images are Restorable: Two-stage Compensation and Alignment Framework for Image Restoration

All-in-One Image Restoration for Unknown Degradations Using Adaptive Discriminative Filters for Specific Degradations

Learning Weather-General and Weather-Specific Features for Image Restoration Under Multiple Adverse Weather Conditions

AccelIR: Task-Aware Image Compression for Accelerating Neural Restoration

Robust Unsupervised StyleGAN Image Restoration

Ingredient-Oriented Multi-Degradation Learning for Image Restoration

Contrastive Semi-supervised Learning for Underwater Image Restoration via Reliable Bank

Nighttime Smartphone Reflective Flare Removal Using Optical Center Symmetry Prior

Robust Single Image Reflection Removal Against Adversarial Attacks

ShadowDiffusion: When Degradation Prior Meets Diffusion Model for Shadow Removal

Document Image Shadow Removal Guided by Color-Aware Background

Generating Aligned Pseudo-Supervision from Non-Aligned Data for Image Restoration in Under-Display Camera

GamutMLP: A Lightweight MLP for Color Loss Recovery

ABCD: Arbitrary Bitwise Coefficient for De-Quantization

Visual Recognition-Driven Image Restoration for Multiple Degradation With Intrinsic Semantics Recovery

Parallel Diffusion Models of Operator and Image for Blind Inverse Problems

Image Reconstruction

Raw Image Reconstruction with Learned Compact Metadata

High-resolution image reconstruction with latent diffusion models from human brain activity

Catch Missing Details: Image Reconstruction with Frequency Augmented Variational Autoencoder

Optimization-Inspired Cross-Attention Transformer for Compressive Sensing

Seeing Beyond the Brain: Conditional Diffusion Model with Sparse Masked Modeling for Vision Decoding

Burst Restoration

Burstormer: Burst Image Restoration and Enhancement Transformer

Gated Multi-Resolution Transfer Network for Burst Restoration and Enhancement

Video Restoration

A Simple Baseline for Video Restoration with Grouped Spatial-temporal Shift

HNeRV: A Hybrid Neural Representation for Videos

Blind Video Deflickering by Neural Filtering with a Flawed Atlas

[Back-to-Overview]

Super Resolution

Image Super Resolution

Activating More Pixels in Image Super-Resolution Transformer

N-Gram in Swin Transformers for Efficient Lightweight Image Super-Resolution

Omni Aggregation Networks for Lightweight Image Super-Resolution

OPE-SR: Orthogonal Position Encoding for Designing a Parameter-free Upsampling Module in Arbitrary-scale Image Super-Resolution

Local Implicit Normalizing Flow for Arbitrary-Scale Image Super-Resolution

Cascaded Local Implicit Transformer for Arbitrary-Scale Super-Resolution

Deep Arbitrary-Scale Image Super-Resolution via Scale-Equivariance Pursuit

CiaoSR: Continuous Implicit Attention-in-Attention Network for Arbitrary-Scale Image Super-Resolution

Super-Resolution Neural Operator

Human Guided Ground-truth Generation for Realistic Image Super-resolution

Better "CMOS" Produces Clearer Images: Learning Space-Variant Blur Estimation for Blind Image Super-Resolution

Implicit Diffusion Models for Continuous Super-Resolution

CABM: Content-Aware Bit Mapping for Single Image Super-Resolution Network with Large Input

Spectral Bayesian Uncertainty for Image Super-Resolution

Cross-Guided Optimization of Radiance Fields With Multi-View Image Super-Resolution for High-Resolution Novel View Synthesis

Image Super-Resolution Using T-Tetromino Pixels

Memory-Friendly Scalable Super-Resolution via Rewinding Lottery Ticket Hypothesis

Equivalent Transformation and Dual Stream Network Construction for Mobile Image Super-Resolution

Perception-Oriented Single Image Super-Resolution using Optimal Objective Estimation

OSRT: Omnidirectional Image Super-Resolution with Distortion-aware Transformer

B-Spline Texture Coefficients Estimator for Screen Content Image Super-Resolution

Spatial-Frequency Mutual Learning for Face Super-Resolution

Learning Generative Structure Prior for Blind Text Image Super-resolution

Guided Depth Super-Resolution by Deep Anisotropic Diffusion

Toward Stable, Interpretable, and Lightweight Hyperspectral Super-Resolution

Zero-Shot Dual-Lens Super-Resolution

Probability-based Global Cross-modal Upsampling for Pansharpening

CutMIB: Boosting Light Field Super-Resolution via Multi-View Image Blending

Quantum Annealing for Single Image Super-Resolution

Bicubic++: Slim, Slimmer, Slimmest -- Designing an Industry-Grade Super-Resolution Network

Hybrid Transformer and CNN Attention Network for Stereo Image Super-resolution

Video Super Resolution

Towards High-Quality and Efficient Video Super-Resolution via Spatial-Temporal Data Overfitting

Structured Sparsity Learning for Efficient Video Super-Resolution

Compression-Aware Video Super-Resolution

Learning Spatial-Temporal Implicit Neural Representations for Event-Guided Video Super-Resolution

Consistent Direct Time-of-Flight Video Depth Super-Resolution

[Back-to-Overview]

Image Rescaling

HyperThumbnail: Real-time 6K Image Rescaling with Rate-distortion Optimization

DINN360: Deformable Invertible Neural Network for Latitude-Aware 360deg Image Rescaling

[Back-to-Overview]

Denoising

Image Denoising

Masked Image Training for Generalizable Deep Image Denoising

Spatially Adaptive Self-Supervised Learning for Real-World Image Denoising

LG-BPN: Local and Global Blind-Patch Network for Self-Supervised Real-World Denoising

Real-time Controllable Denoising for Image and Video

Zero-Shot Noise2Noise: Efficient Image Denoising without any Data

Patch-Craft Self-Supervised Training for Correlated Image Denoising

sRGB Real Noise Synthesizing with Neighboring Correlation-Aware Noise Model

Spectral Enhanced Rectangle Transformer for Hyperspectral Image Denoising

Efficient View Synthesis and 3D-based Multi-Frame Denoising with Multiplane Feature Representations

Structure Aggregation for Cross-Spectral Stereo Image Guided Denoising

Polarized Color Image Denoising

[Back-to-Overview]

Deblurring

Image Deblurring

Structured Kernel Estimation for Photon-Limited Deconvolution

Blur Interpolation Transformer for Real-World Motion from Blur

Neumann Network with Recursive Kernels for Single Image Defocus Deblurring

Efficient Frequency Domain-based Transformers for High-Quality Image Deblurring

Hybrid Neural Rendering for Large-Scale Scenes with Motion Blur

Self-Supervised Non-Uniform Kernel Estimation With Flow-Based Motion Prior for Blind Image Deblurring

Uncertainty-Aware Unsupervised Image Deblurring with Deep Residual Prior

K3DN: Disparity-Aware Kernel Estimation for Dual-Pixel Defocus Deblurring

Self-Supervised Blind Motion Deblurring With Deep Expectation Maximization

HyperCUT: Video Sequence from a Single Blurry Image using Unsupervised Ordering

Video Deblurring

Deep Discriminative Spatial and Temporal Network for Efficient Video Deblurring

[Back-to-Overview]

Deraining

Learning A Sparse Transformer Network for Effective Image Deraining

SmartAssign: Learning a Smart Knowledge Assignment Strategy for Deraining and Desnowing

[Back-to-Overview]

Dehazing

RIDCP: Revitalizing Real Image Dehazing via High-Quality Codebook Priors

Curricular Contrastive Regularization for Physics-aware Single Image Dehazing

Video Dehazing via a Multi-Range Temporal Alignment Network with Physical Prior

SCANet: Self-Paced Semi-Curricular Attention Network for Non-Homogeneous Image Dehazing

Streamlined Global and Local Features Combinator (SGLC) for High Resolution Image Dehazing

[Back-to-Overview]

HDR Imaging / Multi-Exposure Image Fusion

Learning a Practical SDR-to-HDRTV Up-conversion using New Dataset and Degradation Models

SMAE: Few-shot Learning for HDR Deghosting with Saturation-Aware Masked Autoencoders

A Unified HDR Imaging Method with Pixel and Patch Level

Inverting the Imaging Process by Learning an Implicit Camera Model

Joint HDR Denoising and Fusion: A Real-World Mobile HDR Image Dataset

HDR Imaging with Spatially Varying Signal-to-Noise Ratios

1000 FPS HDR Video with a Spike-RGB Hybrid Camera

[Back-to-Overview]

Frame Interpolation

Extracting Motion and Appearance via Inter-Frame Attention for Efficient Video Frame Interpolation

A Unified Pyramid Recurrent Network for Video Frame Interpolation

BiFormer: Learning Bilateral Motion Estimation via Bilateral Transformer for 4K Video Frame Interpolation

AMT: All-Pairs Multi-Field Transforms for Efficient Frame Interpolation

Exploring Discontinuity for Video Frame Interpolation

Frame Interpolation Transformer and Uncertainty Guidance

Exploring Motion Ambiguity and Alignment for High-Quality Video Frame Interpolation

Range-Nullspace Video Frame Interpolation With Focalized Motion Estimation

Event-based Video Frame Interpolation with Cross-Modal Asymmetric Bidirectional Motion Fields

Event-based Blurry Frame Interpolation under Blind Exposure

Event-Based Frame Interpolation with Ad-hoc Deblurring

Joint Video Multi-Frame Interpolation and Deblurring under Unknown Exposure Time

[Back-to-Overview]

Image Enhancement

Realistic Saliency Guided Image Enhancement

Low-Light Image Enhancement

Learning Semantic-Aware Knowledge Guidance for Low-Light Image Enhancement

Visibility Constrained Wide-band Illumination Spectrum Design for Seeing-in-the-Dark

DNF: Decouple and Feedback Network for Seeing in the Dark

You Do Not Need Additional Priors or Regularizers in Retinex-Based Low-Light Image Enhancement

Low-Light Image Enhancement via Structure Modeling and Guidance

Learning a Simple Low-light Image Enhancer from Paired Low-light Instances

[Back-to-Overview]

Image Harmonization/Composition

LEMaRT: Label-Efficient Masked Region Transform for Image Harmonization

Semi-supervised Parametric Real-world Image Harmonization

PCT-Net: Full Resolution Image Harmonization Using Pixel-Wise Color Transformations

ObjectStitch: Object Compositing With Diffusion Model

[Back-to-Overview]

Image Completion/Inpainting

NUWA-LIP: Language-Guided Image Inpainting With Defect-Free VQGAN

Imagen Editor and EditBench: Advancing and Evaluating Text-Guided Image Inpainting

SmartBrush: Text and Shape Guided Object Inpainting with Diffusion Model

Semi-Supervised Video Inpainting with Cycle Consistency Constraints

Deep Stereo Video Inpainting

[Back-to-Overview]

Image Matting

Referring Image Matting

Adaptive Human Matting for Dynamic Videos

Mask-Guided Matting in the Wild

End-to-End Video Matting With Trimap Propagation

Ultrahigh Resolution Image/Video Matting With Spatio-Temporal Sparsity

[Back-to-Overview]

Image Compression

Backdoor Attacks Against Deep Image Compression via Adaptive Frequency Trigger

Context-based Trit-Plane Coding for Progressive Image Compression

Learned Image Compression with Mixed Transformer-CNN Architectures

NVTC: Nonlinear Vector Transform Coding

Multi-Realism Image Compression with a Conditional Generator

LVQAC: Lattice Vector Quantization Coupled with Spatially Adaptive Companding for Efficient Learned Image Compression

Video Compression

Neural Video Compression with Diverse Contexts

Video Compression With Entropy-Constrained Neural Representations

Complexity-Guided Slimmable Decoder for Efficient Deep Video Compression

MMVC: Learned Multi-Mode Video Compression with Block-based Prediction Mode Selection and Density-Adaptive Entropy Coding

Motion Information Propagation for Neural Video Compression

Hierarchical B-Frame Video Coding Using Two-Layer CANF Without Motion Coding

HNeRV: A Hybrid Neural Representation for Videos

[Back-to-Overview]

Image Quality Assessment

Quality-aware Pre-trained Models for Blind Image Quality Assessment

Blind Image Quality Assessment via Vision-Language Correspondence: A Multitask Learning Perspective

Towards Artistic Image Aesthetics Assessment: a Large-scale Dataset and a New Method

Re-IQA: Unsupervised Learning for Image Quality Assessment in the Wild

An Image Quality Assessment Dataset for Portraits

MD-VQA: Multi-Dimensional Quality Assessment for UGC Live Videos

CR-FIQA: Face Image Quality Assessment by Learning Sample Relative Classifiability

SB-VQA: A Stack-Based Video Quality Assessment Framework for Video Enhancement

[Back-to-Overview]

Style Transfer

Fix the Noise: Disentangling Source Feature for Controllable Domain Translation

Neural Preset for Color Style Transfer

CAP-VSTNet: Content Affinity Preserved Versatile Style Transfer

StyleGAN Salon: Multi-View Latent Optimization for Pose-Invariant Hairstyle Transfer

Modernizing Old Photos Using Multiple References via Photorealistic Style Transfer

QuantArt: Quantizing Image Style Transfer Towards High Visual Fidelity

Master: Meta Style Transformer for Controllable Zero-Shot and Few-Shot Artistic Style Transfer

Learning Dynamic Style Kernels for Artistic Style Transfer

Inversion-Based Style Transfer with Diffusion Models

[Back-to-Overview]

Image Editing

Imagic: Text-Based Real Image Editing with Diffusion Models

SINE: SINgle Image Editing with Text-to-Image Diffusion Models

CoralStyleCLIP: Co-optimized Region and Layer Selection for Image Editing

SIEDOB: Semantic Image Editing by Disentangling Object and Background

DiffusionRig: Learning Personalized Priors for Facial Appearance Editing

Paint by Example: Exemplar-based Image Editing with Diffusion Models

StyleRes: Transforming the Residuals for Real Image Editing With StyleGAN

Delving StyleGAN Inversion for Image Editing: A Foundation Latent Space Viewpoint

InstructPix2Pix: Learning to Follow Image Editing Instructions

Deep Curvilinear Editing: Commutative and Nonlinear Image Manipulation for Pretrained Deep Generative Model

Null-text Inversion for Editing Real Images using Guided Diffusion Models

DeltaEdit: Exploring Text-free Training for Text-Driven Image Manipulation

Text-Guided Unsupervised Latent Transformation for Multi-Attribute Image Manipulation

EDICT: Exact Diffusion Inversion via Coupled Transformations

Video Editing

DPE: Disentanglement of Pose and Expression for General Video Portrait Editing

Diffusion Video Autoencoders: Toward Temporally Consistent Face Video Editing via Disentangled Video Encoding

Shape-aware Text-driven Layered Video Editing

[Back-to-Overview]

Image Generation/Synthesis / Image-to-Image Translation

Text-to-Image / Text Guided / Multi-Modal

GALIP: Generative Adversarial CLIPs for Text-to-Image Synthesis

Scaling up GANs for Text-to-Image Synthesis

Variational Distribution Learning for Unsupervised Text-to-Image Generation

Toward Verifiable and Reproducible Human Evaluation for Text-to-Image Generation

Shifted Diffusion for Text-to-image Generation

ReCo: Region-Controlled Text-to-Image Generation

RIATIG: Reliable and Imperceptible Adversarial Text-to-Image Generation With Natural Prompts

GLIGEN: Open-Set Grounded Text-to-Image Generation

Multi-Concept Customization of Text-to-Image Diffusion

ERNIE-ViLG 2.0: Improving Text-to-Image Diffusion Model With Knowledge-Enhanced Mixture-of-Denoising-Experts

Uncovering the Disentanglement Capability in Text-to-Image Diffusion Models

DreamBooth: Fine Tuning Text-to-Image Diffusion Models for Subject-Driven Generation

Specialist Diffusion: Plug-and-Play Sample-Efficient Fine-Tuning of Text-to-Image Diffusion Models To Learn Any Unseen Style

MAGVLT: Masked Generative Vision-and-Language Transformer

Freestyle Layout-to-Image Synthesis

Sound to Visual Scene Generation by Audio-to-Visual Latent Alignment

Collaborative Diffusion for Multi-Modal Face Generation and Editing

SpaText: Spatio-Textual Representation for Controllable Image Generation

Plug-and-Play Diffusion Features for Text-Driven Image-to-Image Translation

LANIT: Language-Driven Image-to-Image Translation for Unlabeled Data

High-Fidelity Guided Image Synthesis with Latent Diffusion Models

Safe Latent Diffusion: Mitigating Inappropriate Degeneration in Diffusion Models

Image-to-Image / Image Guided

Person Image Synthesis via Denoising Diffusion Model

Picture that Sketch: Photorealistic Image Generation from Abstract Sketches

Fine-Grained Face Swapping via Regional GAN Inversion

Masked and Adaptive Transformer for Exemplar Based Image Translation

Zero-shot Generative Model Adaptation via Image-specific Prompt Learning

StyleGene: Crossover and Mutation of Region-Level Facial Genes for Kinship Face Synthesis

Unpaired Image-to-Image Translation With Shortest Path Regularization

BBDM: Image-to-image Translation with Brownian Bridge Diffusion Models

MaskSketch: Unpaired Structure-guided Masked Image Generation

Others for image generation

AdaptiveMix: Improving GAN Training via Feature Space Shrinkage

MAGE: MAsked Generative Encoder to Unify Representation Learning and Image Synthesis

Regularized Vector Quantization for Tokenized Image Synthesis

Exploring Incompatible Knowledge Transfer in Few-shot Image Generation

Post-training Quantization on Diffusion Models

LayoutDiffusion: Controllable Diffusion Model for Layout-to-image Generation

DiffCollage: Parallel Generation of Large Content with Diffusion Models

Few-shot Semantic Image Synthesis with Class Affinity Transfer

NoisyTwins: Class-Consistent and Diverse Image Generation through StyleGANs

DCFace: Synthetic Face Generation with Dual Condition Diffusion Model

Exploring Incompatible Knowledge Transfer in Few-shot Image Generation

Class-Balancing Diffusion Models

Spider GAN: Leveraging Friendly Neighbors to Accelerate GAN Training

Towards Accurate Image Coding: Improved Autoregressive Image Generation with Dynamic Vector Quantization

Not All Image Regions Matter: Masked Vector Quantization for Autoregressive Image Generation

Efficient Scale-Invariant Generator with Column-Row Entangled Pixel Synthesis

Inferring and Leveraging Parts from Object Shape for Improving Semantic Image Synthesis

GLeaD: Improving GANs with A Generator-Leading Task

Where Is My Spot? Few-Shot Image Generation via Latent Subspace Optimization

KD-DLGAN: Data Limited Image Generation via Knowledge Distillation

Private Image Generation With Dual-Purpose Auxiliary Classifier

SceneComposer: Any-Level Semantic Image Synthesis

Exploring Intra-Class Variation Factors With Learnable Cluster Prompts for Semi-Supervised Image Synthesis

Re-GAN: Data-Efficient GANs Training via Architectural Reconfiguration

Discriminator-Cooperated Feature Map Distillation for GAN Compression

Wavelet Diffusion Models are fast and scalable Image Generators

On Distillation of Guided Diffusion Models

Binary Latent Diffusion

All are Worth Words: A ViT Backbone for Diffusion Models

Towards Practical Plug-and-Play Diffusion Models

Lookahead Diffusion Probabilistic Models for Refining Mean Estimation

Diffusion Probabilistic Model Made Slim

Self-Guided Diffusion Models

Video Generation

Conditional Image-to-Video Generation with Latent Flow Diffusion Models

Video Probabilistic Diffusion Models in Projected Latent Space

Decomposed Diffusion Models for High-Quality Video Generation

MoStGAN: Video Generation with Temporal Motion Styles

Align your Latents: High-Resolution Video Synthesis with Latent Diffusion Models

Tell Me What Happened: Unifying Text-guided Video Completion via Multimodal Masked Video Generation

MM-Diffusion: Learning Multi-Modal Diffusion Models for Joint Audio and Video Generation

Dimensionality-Varying Diffusion Process

[Back-to-Overview]

Others

Perspective Fields for Single Image Camera Calibration

DC2: Dual-Camera Defocus Control by Learning to Refocus

Images Speak in Images: A Generalist Painter for In-Context Visual Learning

Make-A-Story: Visual Memory Conditioned Consistent Story Generation

Cross-GAN Auditing: Unsupervised Identification of Attribute Level Similarities and Differences between Pretrained Generative Models

LightPainter: Interactive Portrait Relighting with Freehand Scribble

Neural Texture Synthesis with Guided Correspondence

Uncurated Image-Text Datasets: Shedding Light on Demographic Bias

Large-capacity and Flexible Video Steganography via Invertible Neural Network

Putting People in Their Place: Affordance-Aware Human Insertion into Scenes

Controllable Light Diffusion for Portraits

Talking Head Generation

Seeing What You Said: Talking Face Generation Guided by a Lip Reading Expert

High-Fidelity and Freely Controllable Talking Head Video Generation

MetaPortrait: Identity-Preserving Talking Head Generation with Fast Personalized Adaptation

Identity-Preserving Talking Face Generation with Landmark and Appearance Priors

LipFormer: High-Fidelity and Generalizable Talking Face Generation With a Pre-Learned Facial Codebook

High-fidelity Generalized Emotional Talking Face Generation with Multi-modal Emotion Space Learning

DiffTalk: Crafting Diffusion Models for Generalized Audio-Driven Portraits Animation

Virtual Try-on

GP-VTON: Towards General Purpose Virtual Try-on via Collaborative Local-Flow Global-Parsing Learning

Linking Garment With Person via Semantically Associated Landmarks for Virtual Try-On

TryOnDiffusion: A Tale of Two UNets

Handwriting/Font Generation

CF-Font: Content Fusion for Few-shot Font Generation

Neural Transformation Fields for Arbitrary-Styled Font Generation

DeepVecFont-v2: Exploiting Transformers to Synthesize Vector Fonts with Higher Quality

Handwritten Text Generation from Visual Archetypes

Disentangling Writer and Character Styles for Handwriting Generation

Conditional Text Image Generation With Diffusion Models

Layout Generation

Unifying Layout Generation with a Decoupled Diffusion Model

Unsupervised Domain Adaption with Pixel-level Discriminator for Image-aware Layout Generation

PosterLayout: A New Benchmark and Approach for Content-aware Visual-Textual Presentation Layout

LayoutDM: Discrete Diffusion Model for Controllable Layout Generation

LayoutDM: Transformer-based Diffusion Model for Layout Generation

About

A Collection of Papers and Codes in CVPR2023/2022 about low level vision