Mayank022's picture
Update README.md
5a1f194 verified
metadata
license: mit

facegen-facenet-unet-gan-embedding

Model Overview

image/png

This repository hosts a generative model trained to synthesize 128×128 face images conditioned on facial embeddings. The architecture jointly trains a FaceNet encoder and a UNet-based GAN generator to produce high-fidelity images from identity embeddings.

Codebase, Dataset, and Model Artifacts

Category Description Link
GitHub Repo Full training and inference codebase GitHub Repository
Dataset Cropped faces (128x128) for training Hugging Face Dataset
Trained Model Final GAN model with FaceNet encoder Hugging Face Model
Training Notebook End-to-end model training pipeline in Colab Colab Notebook
Inference Notebook Generate images from embeddings Colab Notebook

Experimental Summary

Find detailed metrics, loss trends, and inference samples in the full Weights & Biases report:

Weights & Biases Report – Training, Metrics, and Sample Outputs

Use Case

The model can be used for:

  • Face reconstruction from embeddings
  • Conditional face generation
  • Evaluating zero-shot generalization for unseen identities

Architecture

  • Encoder: Pretrained FaceNet (jointly fine-tuned)
  • Generator: Modified UNet with residual upsampling blocks and conditional embedding injection at multiple resolutions
  • Embedding Injection: Done at 8×8 resolution with skip connections

Training Details

  • Training Time: ~6 hours on A100
  • Dataset: 12,000 human face images
  • Losses: MSE + Perceptual + Adversarial
  • Optimizers: AdamW with learning rate scheduling

Checkpoints

  • epoch_100/
  • epoch_200/
  • epoch_300/ (final)

Intended Use

This model is intended for research purposes, particularly for understanding conditional generation and face embedding interpretability.

Limitations

  • Trained on a limited dataset
  • May not generalize well to non-human faces or distorted embeddings