readme.md

May 10, 2024 ยท View on GitHub

neosr

neosr is a framework for training real-world single-image super-resolution networks. wiki.


Join our Discord
news 05.09.2024 - Release Real-PLKSR network. wiki
05.08.2024 - Release Nomos-v2 dataset. For more details, see datasets

๐Ÿค support me

Tip

Consider supporting me on KoFi โ˜• or Patreon

๐Ÿ’ป installation

Requires Python 3.11 and CUDA =>11.8. Install latest Pytorch (=>2.1) and TorchVision (required).

Clone the repository:

git clone https://github.com/muslll/neosr
cd neosr

Then install other dependencies via pip:

pip install -e .

Alternatively, use poetry (recommended on linux):

poetry install
poetry add torch@latest torchvision@latest

Note: You must use poetry shell to enter the env after installation.

(optional) If you want to convert your models (convert.py), you need the following dependencies:

pip install onnx onnxruntime-gpu onnxconverter-common onnxsim

You can also install using poetry (recommended on linux):

poetry add onnx onnxruntime-gpu onnxconverter-common onnxsim

Please read the wiki tutorial for converting your models.

quick start

Start training by running:

python train.py -opt options.yml

Where options.yml is a configuration file. Templates can be found in options.

Tip

Please read the wiki Configuration Walkthrough for an explanation of each option.

features

Supported archs:

archoption
Real-ESRGANesrgan
SRVGGNetCompactcompact
SwinIRswinir_small, swinir_medium
HAThat_s, hat_m, hat_l
OmniSRomnisr
SRFormersrformer_light, srformer_medium
DATdat_small, dat_medium, dat_2
DITNditn
DCTLSAdctlsa
SPANspan
Real-CUGANcugan
CRAFTcraft
SAFMNsafmn, safmn_l
RGTrgt, rgt_s
ATDatd, atd_light
PLKSRplksr, plksr_tiny
RealPLKSRrealplksr, realplksr_s
DRCTdrct, drct_l, drct_s
EFENefen

Note

For all arch-specific parameters, read the wiki.

Under Testing

archoption

Supported Discriminators:

netoption
U-Net w/ SNunet
A2-FPN w/ SNa2fpn
PatchGAN w/ SNpatchgan

Supported Optimizers:

optimizeroption
AdamAdam or adam
AdamWAdamW or adamw
NAdamNAdam or nadam
LionLion or lion
LAMBLamb or lamb
AdanAdan or adan

Supported losses:

lossoption
L1 LossL1Loss, l1
L2 LossMSELoss, l2
Huber LossHuberLoss, huber
CHC (Clipped Huber with Cosine Similarity Loss)chc, chc_l2
Perceptual Lossperceptual_opt, PerceptualLoss
GANgan_opt, GANLoss, MultiScaleGANLoss
YCbCr Color Loss (bt601)color_opt, colorloss
Luma Loss (CIE L*)luma_opt lumaloss
MS-SSIMmssim_opt mssim
LDL Lossldl_opt
Focal Frequencyff_opt, focalfrequencyloss
DISTSdists_opt, dists
Wavelet Guidedwavelet_guided
Gradient-Weightedgw_opt, gw_loss

Supported Augmentations:

augmentationoption
Rotationuse_rot
Flipuse_hflip
MixUpmixup
CutMixcutmix
ResizeMixresizemix
CutBlurcutblur

Supported models:

modeldescriptionoption
DefaultBase model, supports both Generator and Discriminatordefault
OTFBuilds on top of default, adding Real-ESRGAN on-the-fly degradationsotf

Supported dataset loaders:

loaderoption
Paired datasetspaired
Single datasets (for inference, no GT required)single
Real-ESRGAN on-the-fly degradationotf

๐Ÿ“ธ datasets

As part of neosr, I have released a dataset series called Nomos. The purpose of these datasets is to distill only the best images from the academic and community datasets. A total of 14 datasets were manually reviewed and processed, including: Adobe-MIT-5k, RAISE, LSDIR, LIU4k-v2, KONIQ-10k, Nikon LL RAW, DIV8k, FFHQ, Flickr2k, ModernAnimation1080_v2, Rawsamples, SignatureEdits, Hasselblad raw samples and Unsplash.

  • Nomos-v2 (recommended): contains 6000 images, multipurpose. Data distribution:
pie
  title Nomos-v2 distribution
  "Animal / fur" : 439
  "Interiors" : 280
  "Exteriors / misc" : 696
  "Architecture / geometric" : 1470
  "Drawing / painting / anime" : 1076
  "Humans" : 598
  "Mountain / Rocks" : 317
  "Text" : 102
  "Textures" : 439
  "Vegetation" : 574
  • nomos_uni (recommended for lightweight networks): contains 2989 images, multipurpose. Meant to be used on lightweight networks (<800k parameters).
  • hfa2k: contains 2568 anime images.
dataset downloadsha256
nomosv2 (3GB)sha256
nomosv2.lmdb (3GB)sha256
nomosv2_lq_4x (187MB)sha256
nomosv2_lq_4x.lmdb (187MB)sha256
nomos_uni (1.3GB)sha256
nomos_uni.lmdb (1.3GB)sha256
nomos_uni_lq_4xsha256
nomos_uni_lq_4x.lmdbsha256
hfa2ksha256

community datasets

Datasets made by the upscaling community. More info can be found in the Enhance Everything discord

  • FaceUp: Curated version of FFHQ
  • SSDIR: Curated version of LSDIR.
  • kim's 8k Dataset V2: Video Game Dataset
datasetdownload
@Kim2091 8k Dataset V2GDrive (33.5GB)
@Phhofm FaceUpGDrive (4GB)
@Phhofm SSDIRGdrive (4.5GB)

resources

๐Ÿ“„ license and acknowledgements

Released under the Apache license. All licenses listed on license/readme. This code was originally based on BasicSR.

Thanks to victorca25/traiNNer, styler00dollar/Colab-traiNNer and timm for providing helpful insights into some problems.

Thanks to contributors @Phhofm, @Sirosky, @Kim2091, @terrainer, @Corpsecreate and @umzi2 for helping with tests and bug reporting.