MossNet
MossNet is a neural network for solving super resolution problem(SR - further) preserving memory efficiency.
Abstract
The current state-of-the-art super resolution methods are based on diffusion models, which are computationally expensive. In the same time also much model utilize text-to-image models which are making inference even more complex. The MossNet aims to remove this complexity by utilizing U-Net architecture without text-to-image model.
Architecture
MossNet proposes context-aware patch-local upscaling mechanism, which allow to use single embedding "describing" image for whole upscaling process. In the same using smaller, fixed patches allow to reduce memory footprint.
Preliminary results
Experiment setup
Model training did not include noise reduction targets or deblur targets. The reason behind this choice is that the model is aimed to users of handsets and embedded devices that may want blurred images to be displayed as is and denoise algorithms are usually separate. Thus metrics are limited to LPIPS and PSNR. The model was trained of 50000 iterations on a open images datasets. The model then was distilled to XS version.
The model was evaluated on OpenImages dataset which was also used for training. Also to assess the model out-of-sample generalization ability arbitrary images from DIV2K and Unsplash were used during visual testing.
The GFLOPS parameter was measured on SR for image of 256x256 pixels.
Quantative results
Model | LPIPS↓ | PSNR↑ | GFLOPS | Param Count |
---|---|---|---|---|
MossNet-XS | 0.132 | 27.64 | 3.32 | 64.2K |
SRGAN⚠️ | 0.126 | 26.63 | - | 5.949M |
StableSR@200 steps | 0.311 | 23.26 | - | approx. 200M |
GuideSR | 0.265 | 24.76 | - | approx. 200M |
EDSR⚠️ | 0.133 | 34.64 | - | 1.37M |
Important notice: The above quantative results may have bias since the model was never trained on Div2K dataset, unlike models from this this paper. Models for which data is likely affected by this bias are marked by a warning⚠️ icon.
Though this issue is awaiting model(among others) to be specifically re-trained on DIV2K, the problem with my GPU setup would like take time to resolve. For the same reason GFLOPS is currently unavailable for most models in above table.
Visual results
DIV2K
OpenImages
Unsplash
Just random image from Unsplash.References
- Park, S.‑H., Moon, Y.‑S., & Cho, N. I. (2022). Perception‑Oriented Single Image Super‑Resolution using Optimal Objective Estimation. arXiv:2211.13676.
- Park, S.‑H., Moon, Y.‑S., & Cho, N. I. (2023). Perception‑Oriented Single Image Super‑Resolution Using Optimal Objective Estimation. In CVPR 2023.
- Wang, J., Yue, Z., Zhou, S., Chan, K. C. K., & Loy, C. C. (2023). Exploiting Diffusion Prior for Real‑World Image Super‑Resolution. arXiv:2305.07015.
- Arora, A., Tu, Z., Wang, Y., Bai, R., Wang, J., & Ma, S. (2025). GuideSR: Rethinking Guidance for One‑Step High‑Fidelity Diffusion‑Based Super‑Resolution. arXiv:2505.00687.
- Hadji et al. (2025). Edge‑SD‑SR: Low Latency and Parameter Efficient On‑device Super‑Resolution with Stable Diffusion. In CVPR 2025.
- Li, H., Yang, Y., Chang, M., Feng, H., Xu, Z., Li, Q., & Chen, Y. (2021). SRDiff: Single Image Super‑Resolution with Diffusion Probabilistic Models. arXiv:2104.14951.
- Dietz, T., Moser, B. B., Nauen, T., Raue, F., Frolov, S., & Dengel, A. (2025). A Study in Dataset Distillation for Image Super‑Resolution. arXiv:2502.03656.
- Agustsson, E., & Timofte, R. (2017). NTIRE 2017 Challenge on Single Image Super‑Resolution: Dataset and Study. In CVPR Workshops.
- Agustsson, E., & Timofte, R. (2018). DIV2K – Diverse 2K resolution image dataset (800 train + 100 val + 100 test).
- Lim, B., Son, S., Kim, H., Nah, S., & Lee, K. M. (2017). Flickr2K – 2,650 high-resolution images for SR, used in EDSR. (Collected via Flickr API)
- Kuznetsova, A., Rom, H., Alldrin, N., Uijlings, J., Krasin, I., Pont‑Tuset, J., et al. (2018). The Open Images Dataset V4. arXiv:1811.00982.
Notice
As this research is still in progress, this page is left under restrictive CC license.Copyright
This page is licensed under the CC BY-NC-ND license.
Author

Flora
Rust-loving, Python-purring Rubyist with a taste for clean UI and warm naps in the sun.