GAN-based CelebA Face Generation: Utilized GAN architecture to generate realistic celebrity face images from the CelebA dataset, showcasing benefits such as unsupervised learning and diverse data distribution handling, while acknowledging trade-offs like computational overhead and stability issues.
| Name | College Roll No. | NM ID | |
|---|---|---|---|
| Fowzaan | fowzaan.rasheed@gmail.com | 2021503017 | 8E4AF1FB4D2CAD089814D6BED938AC27 |
This project implements a Generative Adversarial Network (GAN) to generate realistic celebrity face images using the CelebA dataset. GANs are a type of neural network architecture consisting of two networks, a Generator and a Discriminator, which are trained simultaneously to produce high-quality synthetic data that is indistinguishable from real data.
This project aims to generate realistic celebrity face images using GANs. It follows the standard GAN architecture, where the Generator network generates synthetic images, and the Discriminator network distinguishes between real and fake images. The two networks are trained adversarially until the Generator produces images that are indistinguishable from real images.
- Python 3.x
- PyTorch
- torchvision
- matplotlib
- numpy
-
Clone the repository:
git clone https://github.com/your-username/celeba-face-generation.git
-
Download the CelebA dataset from here and place it in the
./datasetdirectory. -
Install the required Python dependencies:
pip install -r requirements.txt
-
Run the training script:
python train.py
The trained GAN model produces high-quality celebrity face images that are visually similar to real images. The results can be found in the ./results directory.
- GANs are an unsupervised learning method, eliminating the need for manually labeled data.
- GANs generate examples from a given dataset, increasing efficiency by producing realistic results.
- Discriminator, when trained, can work as a classifier.
- GANs can learn complicated and messy data distributions, tackling diverse problems.
- GANs can be adapted to multiple formats like text, video, image, audio, etc.
- Requires significant computational overhead.
- Generated results may sometimes be distinguishable from real data.
- Generating high-quality results from text and audio is challenging.
- Training can be unstable between Generator and Discriminator.
- Lack of density estimation and intrinsic evaluation metrics.
Contributions are welcome! Feel free to open issues or submit pull requests for any improvements or additional features.
This project is licensed under the MIT License - see the LICENSE file for details.
