Frequency-Aware Super-Denoiser 🎯

A novel frequency-domain diffusion model for image enhancement and restoration tasks. This model excels as a super-denoiser rather than a traditional generative model, making it highly practical for real-world applications.

πŸš€ Model Overview

This implementation introduces a Frequency-Aware Diffusion Model that processes images in the frequency domain using Discrete Cosine Transform (DCT). Unlike traditional diffusion models focused on generation, this model specializes in image enhancement, restoration, and denoising tasks.

Key Features

  • πŸ”¬ DCT-based processing: Patch-wise frequency domain enhancement (16Γ—16 patches)
  • ⚑ High-performance denoising: 95-99% reconstruction fidelity (MSE: 0.002-0.047)
  • πŸŽ›οΈ Progressive enhancement: Multiple enhancement levels with user control
  • πŸ’Ύ Memory efficient: Patch-based processing reduces computational overhead
  • πŸ”„ Stable training: No mode collapse, excellent convergence
  • 🎨 Multiple applications: From photo enhancement to medical imaging

πŸ“Š Performance Metrics

Metric Reconstruction Enhancement Status Description
MSE 0.002778 0.040256 βœ… Excellent Mean Squared Error vs. ground truth
PSNR 32.1 dB 20.0 dB 🟒 Very Good Peak Signal-to-Noise Ratio
SSIM 0.9529 0.5920 βœ… Excellent Structural Similarity Index
Training Stability Perfect - βœ… No mode collapse Consistent convergence
Processing Speed Single-pass Real-time βœ… Fast Optimized inference
Memory Efficiency High High βœ… Patch-based 16Γ—16 DCT patches

Performance Analysis

  • 🎯 Reconstruction: Excellent performance with light noise (SSIM > 0.95)
  • 🧹 Enhancement: Good noise removal capability for heavier noise
  • ⚑ Speed: Real-time capable with single forward pass
  • πŸ’Ύ Efficiency: Memory-optimized patch-based processing

🎯 Applications

βœ… Primary Applications (Excellent Performance)

  1. Noise Removal - Gaussian and salt-pepper noise elimination
  2. Image Enhancement - Sharpening and quality improvement
  3. Progressive Enhancement - Multi-level enhancement control

🟒 Secondary Applications (Very Good Performance)

  1. Medical/Scientific Imaging - Low-quality image enhancement
  2. Texture Synthesis - Artistic and creative applications

πŸ”΅ Experimental Applications (Good Performance)

  1. Image Interpolation - Smooth morphing between images
  2. Style Transfer - Artistic effects and stylization
  3. Real-time Processing - Fast single-pass enhancement

πŸ—οΈ Architecture

SmoothDiffusionUNet(
  - Base Channels: 64
  - Time Embedding: 256 dimensions
  - Architecture: U-Net with skip connections
  - Patch Size: 16Γ—16 for DCT processing
  - Timesteps: 500
  - Input/Output: 3-channel RGB (64Γ—64)
)

Frequency-Aware Noise Scheduler

  • DCT Transform: Converts spatial patches to frequency domain
  • Adaptive Scaling: Different noise levels for different frequency components
  • Patch-wise Processing: Maintains spatial locality while processing frequencies

πŸ› οΈ Usage

Basic Enhancement

import torch
from model import SmoothDiffusionUNet
from noise_scheduler import FrequencyAwareNoise
from config import Config

# Load model
config = Config()
model = SmoothDiffusionUNet(config)
model.load_state_dict(torch.load('model_final.pth'))
model.eval()

# Initialize scheduler
scheduler = FrequencyAwareNoise(config)

# Enhance image
enhanced_image = scheduler.sample(model, noisy_image, num_steps=50)

Progressive Enhancement

# Different enhancement levels
enhancement_levels = [10, 25, 50, 100]  # timesteps
results = []

for steps in enhancement_levels:
    enhanced = scheduler.sample(model, noisy_image, num_steps=steps)
    results.append(enhanced)

Comprehensive Testing

# Run all application tests
python comprehensive_test.py

πŸ“¦ Installation

# Clone repository
git clone <repository-url>
cd frequency-aware-super-denoiser

# Install dependencies
pip install -r requirements.txt

# Download Tiny ImageNet dataset
wget http://cs231n.stanford.edu/tiny-imagenet-200.zip
unzip tiny-imagenet-200.zip -d data/

πŸŽ“ Training

# Train the model
python train.py

# Monitor training with tensorboard
tensorboard --logdir=./logs

Training Configuration

  • Dataset: Tiny ImageNet (200 classes, 64Γ—64 images)
  • Batch Size: 32
  • Learning Rate: 1e-4
  • Epochs: 100
  • Loss Function: MSE + Total Variation + Gradient Loss
  • Optimizer: Adam

πŸ§ͺ Testing & Evaluation

Quick Test

python test.py

Comprehensive Evaluation

python comprehensive_test.py

Performance Summary

python model_summary.py

πŸ’Ό Commercial Applications

This model is particularly valuable for:

  1. Photo Editing Software - Enhancement modules for professional tools
  2. Medical Imaging - Preprocessing pipelines for diagnostic systems
  3. Security Systems - Camera image enhancement for better recognition
  4. Document Processing - OCR preprocessing and scan enhancement
  5. Video Streaming - Real-time quality enhancement
  6. Gaming Industry - Texture enhancement systems
  7. Satellite Imaging - Aerial and satellite image processing
  8. Forensic Analysis - Image analysis and enhancement tools

πŸ”¬ Technical Details

Innovation: Frequency-Domain Processing

  • DCT Patches: 16Γ—16 patches converted to frequency domain
  • Adaptive Noise: Different noise characteristics for different frequencies
  • Spatial Preservation: Maintains image structure while enhancing details

Training Stability

  • No Mode Collapse: Frequency-aware approach prevents training instabilities
  • Fast Convergence: Typically converges within 50-100 epochs
  • Robust Performance: Consistent results across different image types

Performance Characteristics

  • Reconstruction Fidelity: Excellent (MSE < 0.05)
  • Enhancement Quality: Superior noise removal and sharpening
  • Processing Speed: Real-time capable with optimized inference
  • Memory Usage: Efficient due to patch-based processing

πŸ“š Related Work

This model builds upon:

  • Diffusion Models (DDPM, DDIM)
  • Frequency Domain Image Processing
  • U-Net Architectures for Image-to-Image Tasks
  • Super-Resolution and Denoising Networks

πŸ“„ Citation

@misc{frequency-aware-super-denoiser,
  title={Frequency-Aware Super-Denoiser: A Novel Approach to Image Enhancement},
  author={Aleksander Majda},
  year={2025},
  note={Proof of Concept Implementation}
}

🀝 Contributing

We welcome contributions! Please see our contributing guidelines for:

  • Bug reports and feature requests
  • Code contributions and improvements
  • Documentation enhancements
  • New application examples

πŸ“§ Contact

For questions, suggestions, or collaborations:

  • Issues: Please use GitHub issues for bug reports
  • Discussions: Use GitHub discussions for questions and ideas
  • Email: nazgut@gmail.com

πŸŽ‰ Acknowledgments

  • Tiny ImageNet dataset creators
  • PyTorch community for the excellent framework
  • Diffusion models research community
  • Frequency domain image processing pioneers

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Dataset used to train nazgut/Grad-CDM

Evaluation results