CT to MRI Translation Using CycleGAN

Project Type: Deep Learning ยท Image Translation

Overview

This project applies a CycleGAN-based architecture to translate CT scans into MRI images and vice versa. The model learns domain-specific features from each modality and performs unpaired image-to-image translation using cycle consistency. This enables synthetic generation of one modality from the other without requiring aligned datasets.

Why This Matters

Medical imaging modalities like CT and MRI provide complementary views of anatomy, but acquiring both can be expensive, time-consuming, or contraindicated. A tool that can generate one from the other helps in clinical decision support, data augmentation, or low-resource settings where one modality is unavailable.

How It Works

The system uses two generator-discriminator pairs trained adversarially. Cycle consistency ensures that an image translated from CT to MRI and back remains faithful to the original. This method allows the model to retain anatomical structure while adapting texture, brightness, and contrast between domains.

Technologies Used

Python, PyTorch, CycleGAN, NumPy, OpenCV, Matplotlib

CycleGAN CT to MRI example

๐Ÿ”— View GitHub Repository

โ† Back to Portfolio