You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

MedVAE

Variational Autoencoders (VAEs) are widely used in imaging tasks such as image generation, reconstruction, and representation learning. However, most open-source VAEs are trained on natural images and are not suitable for 3D medical CT data. This creates a gap for researchers who need reliable VAE models that can handle real clinical CT volumes.

To address this problem, we benchmark MedVAE, a VAE pre-trained on large-scale 3D medical CT data. The goal is to evaluate whether MedVAE can better preserve anatomical structures and intensity distributions compared to general-purpose VAEs.

Benchmarks

Name VAE Type # Patients MAE SSIM DSC DSC Abdominal
stable-diffusion-v1-5 KL-VAE 0
stable-diffusion-3.5-large KL-VAE 0
MedVAE_KL-smooth-tianyu-2K KL-VAE 2k
MedVAE_KL-smooth-kuma-10K KL-VAE 10k
MedVAE_KL-sharp-kuma-10K KL-VAE 10k
MedVAE_KL-sharp-kuma-20K KL-VAE 20k
MedVAE_KL-sharp-kuma-100K KL-VAE 100k
MedVAE_KL-sharp-kuma-300K (ongoing) KL-VAE 300k
MedVAE_KL-sharp-kuma-1M (ongoing) KL-VAE 1M
MedVAE_KL-sharp-kuma-10M (ongoing) KL-VAE 10M

Citation

@article{liu2025see,
  title={See More, Change Less: Anatomy-Aware Diffusion for Contrast Enhancement},
  author={Liu, Junqi and Wu, Zejun and Bassi, Pedro RAS and Zhou, Xinze and Li, Wenxuan and Hamamci, Ibrahim E and Er, Sezgin and Lin, Tianyu and Luo, Yi and Płotka, Szymon and others},
  journal={arXiv preprint arXiv:https://www.arxiv.org/abs/2512.07251},
  year={2025},
  url={https://github.com/MrGiovanni/SMILE}
}
Downloads last month
27
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for SMILE-project/MedVAE

Finetuned
(325)
this model

Paper for SMILE-project/MedVAE