This project focuses on the implementation of a Conditional Generative Adversarial Network (cGAN) for image-to-image translation, specifically targeting the generation of realistic segmented images from input images.
The primary objective is to make image segmentation more accessible while providing valuable experience in utilizing cGANs.
The project uses a subset of the Cityscapes dataset, a widely used dataset for semantic segmentation.
For evaluation, pixel-wise accuracy is used to measure the percentage of correctly classified pixels in the generated segmentation compared to the ground truth. This metric provides insights into the model's performance on the validation set.
In addition to the cGAN, a U-Net model is implemented and trained for comparison. The U-Net architecture, known for its effectiveness in image segmentation, is used to assess the cGAN's performance.
The project compares the results of the cGAN and U-Net models, providing insights into their respective strengths and weaknesses.
Pixel-wise accuracy on the validation set is calculated to quantify the performance of both models.
Below are showed the images generated after 50 epochs on the training set