HamadYA / GhostFaceNets

This repository contains the official implementation of GhostFaceNets, State-Of-The-Art lightweight face recognition models.

Home Page:https://ieeexplore.ieee.org/document/10098610

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

GhostFaceNets

Introduction

  • This repository contains the official implementation of GhostFaceNets, State-Of-The-Art lightweight face recognition models.
  • Accepted in IEEE Access
  • Paper link

GhostFaceNets

Other Implementations

Results

  • IJBB and IJBC are scored at TAR@FAR=1e-4

  • Links in Model backbone are h5 models in Google drive. GhostFaceNet-w-s (loss) where w refers to width, s refers to strides, and loss refers to the loss function {A refers to ArcFace, C refers to CosFace, and SCA refers to Subcenter ArcFace}.

  • In my last experiments, I was able to get 99.7833% and 99.7167% accuracies on LFW using GhostFaceNetV1 S1 and S2 trained on MS1MV3 dataset. To reproduce, simply download lfw.bin, follow the Datasets and Data Preparation, download the weights, and then modify and run evaluation.ipynb. Note: Set flip=True inside the code.

  • GhostFaceNetV1

Model Training Dataset lfw (%) cfp_fp (%) agedb_30 (%) IJBB IJBC
GhostFaceNetV1-0.5-2 (A) MS1MV2 99.783 83.4 91.13 - -
GhostFaceNetV1-1-2 (A) MS1MV2 99.383 88.81 94.85 - -
GhostFaceNetV1-1.2-2 (A) MS1MV2 99.53 90.13 95.62 - -
GhostFaceNetV1-1.4-2 (A) MS1MV2 99.6167 90.8857 95.78 - -
GhostFaceNetV1-1.3-2 (A) MS1MV2 99.65 91.51 96.18 90.53 92.66
GhostFaceNetV1-1.3-1 (A) MS1MV2 99.7667 95.64 97.58 92.19 94.06
GhostFaceNetV1-1.3-2 (C) MS1MV2 99.5167 90.37 95.23 - -
GhostFaceNetV1-1.3-2 (SCA) MS1MV2 99.633 90.73 95.6 - -
GhostFaceNetV1-1.3-1 (C) MS1MV2 99.7167 94.7429 97.1833 - -
GhostFaceNetV1-1.3-1 (SCA) MS1MV2 99.667 91.46 95.85 - -
GhostFaceNetV1-1.3-1 (A) MS1MV3 99.73 96.83 98 93.12 94.94
GhostFaceNetV1-1.3-2 (A) MS1MV3 99.68 93.31 96.92 91.25 93.45
GhostFaceNetV1-2-1 (A) MS1MV2 99.77 96.56 97.47 - -
  • GhostFaceNetV2

    Model Training Dataset lfw (%) cfp_fp (%) agedb_30 (%) IJBB IJBC
    GhostFaceNetV2-1-2 (A) MS1MV2 99.65 92.9 96.212 - -
    GhostFaceNetV2-1.1-2 (A) MS1MV2 99.6 91.814 95.683 - -
    GhostFaceNetV2-1.2-2 (A) MS1MV2 99.633 92.686 95.9 - -
    GhostFaceNetV2-1.4-2 (A) MS1MV2 99.683 91.986 95.983 - -
    GhostFaceNetV2-1.5-2 (A) MS1MV2 99.7 93.9 96.483 - -
    GhostFaceNetV2-1-2 (A) MS1MV3 99.667 94.857 96.417 - -
    GhostFaceNetV2-1.1-2 (A) MS1MV3 99.667 92.471 96.167 - -
    GhostFaceNetV2-1.2-2 (A) MS1MV3 99.683 94.343 96.383 - -
    GhostFaceNetV2-1.4-2 (A) MS1MV3 99.683 94.417 96.417 - -
    GhostFaceNetV2-1.5-2 (A) MS1MV3 99.676 94.857 96.417 - -
    GhostFaceNetV2-1.3-2 (A) MS1MV2 99.717 93.071 96.55 91.7624
    GhostFaceNetV2-1.3-2 (A) MS1MV3 99.683 94.2857 96.833 91.889
    GhostFaceNetV2-1.3-2 (C) MS1MV2 99.584 91.931 95.6 - -
    GhostFaceNetV2-1.3-2 (SCA) MS1MV2 99.7 92.291 95.97 - -
    GhostFaceNetV2-1.3-1 (C) MS1MV2 99.7167 94.7429 97.1833 - -
    GhostFaceNetV2-1.3-1 (SCA) MS1MV2 99.667 91.46 95.85 - -
    GhostFaceNetV2-1.3-1 (A) MS1MV3 99.73 96.83 98 93.12 94.94
    GhostFaceNetV2-1.3-2 (A) MS1MV3 99.68 93.31 96.92 91.25 93.45
    GhostFaceNetV2-1.3-1 (A) MS1MV2 - -
    GhostFaceNetV2-1.3-1 (A) MS1MV3 - -

Installation

To run GhostFaceNets, please install the following packages:

1. Python 3.9.12 64-bit
2. TensorFlow 2.8.0 or above (CUDA compatible GPU needed for GPU training)
3. Keras 2.8.0 or above
4. keras_cv_attention_models
5. glob2
6. pandas
7. tqdm
8. sklearn
9. scikit-image

Datasets and Data Preparation

  • The training and testing datasets (LFW CFP-FP AgeDB-30 included inside the training datasets) can be downloaded from the following URLs:

    1. MS1M-ArcFace - MS1MV2
    2. MS1M-RetinaFace - MS1MV3
  • Download and extract the training dataset in the datasets directory.

  • The extracted folder contains .bin files of the training and testing datasets.

  • To convert .bin files of the training and testing datasets, please run the following commands:

  • prepare_data.py script, Extract data from mxnet record format to folders.

    # Convert `datasets/faces_emore` to `datasets/faces_emore_112x112_folders`
    python prepare_data.py -D datasets/faces_emore
    
    # Convert evaluating bin files
    python prepare_data.py -D datasets/faces_emore -T lfw.bin cfp_fp.bin agedb_30.bin
  • Training dataset Required is a folder including person folders, each person folder including multi face images. Format like

    .               # dataset folder
    ├── 0           # person folder
    │   ├── 100.jpg # face image
    │   ├── 101.jpg # face image
    │   └── 102.jpg # face image
    ├── 1           # person folder
    │   ├── 111.jpg
    │   ├── 112.jpg
    │   └── 113.jpg
    ├── 10
    │   ├── 707.jpg
    │   ├── 708.jpg
    │   └── 709.jpg
  • Evaluating bin files include jpeg image data pairs, and a label indicating if it's a same person, so there are double images than labels

    #    bins   | issame_list
    img_1 img_2 | True
    img_3 img_4 | True
    img_5 img_6 | False
    img_7 img_8 | False

    Image data in bin files like CFP-FP AgeDB-30 is not compatible with tf.image.decode_jpeg, we need to reformat it, which is done by -T parameter.

    ''' Throw error if not reformated yet '''
    ValueError: Can't convert non-rectangular Python sequence to Tensor.

Project Structure

  • Model Modules
  • Training Modules
    • train.py contains a Train class. It uses a scheduler to connect different loss / optimizer / epochs. The basic function is simply basic_model --> build dataset --> add output layer --> add callbacks --> compile --> fit.
    • data.py loads image data as tf.dataset for training. Triplet dataset is different from others.
    • evals.py contains evaluating callback using bin files.
    • losses.py contains arcface / cosface loss functions.
    • myCallbacks.py contains my other callbacks, like saving model / learning rate adjusting / save history.
    • IJB_evals.py evaluates model accuracy using insightface/evaluation/IJB/ datasets.
    • eval_folder.py Run model evaluation on any custom dataset folder, which is in the same format with Training dataset.
    • augment.py including implementation of RandAug and AutoAug.
  • Other Modules

Basic Training

  • We included all training scripts in the paper in training_scripts.ipynb

  • Training example train.Train is mostly functioned as a scheduler.

    from tensorflow import keras
    import losses, train, GhostFaceNets
    
    data_path = 'datasets/faces_emore_112x112_folders'
    eval_paths = ['datasets/faces_emore/lfw.bin', 'datasets/faces_emore/cfp_fp.bin', 'datasets/faces_emore/agedb_30.bin']
    
    basic_model = GhostFaceNets.buildin_models("ghostnetv1", dropout=0, emb_shape=512, output_layer='GDC', bn_momentum=0.9, bn_epsilon=1e-5)
    basic_model = GhostFaceNets.add_l2_regularizer_2_model(basic_model, weight_decay=5e-4, apply_to_batch_normal=False)
    basic_model = GhostFaceNets.replace_ReLU_with_PReLU(basic_model)
    
    tt = train.Train(data_path, eval_paths=eval_paths,
    save_path='ghostnetv1_w1.3_s2.h5', basic_model=basic_model, model=None, 
    lr_base=0.1, lr_decay=0.5, lr_decay_steps=45, lr_min=1e-5,
    batch_size=128, random_status=0, eval_freq=1, output_weight_decay=1)
    
    optimizer = keras.optimizers.SGD(learning_rate=0.1, momentum=0.9)
    sch = [
        {"loss": losses.ArcfaceLoss(scale=32), "epoch": 1, "optimizer": optimizer},
        {"loss": losses.ArcfaceLoss(scale=64), "epoch": 50},
    ]
    tt.train(sch, 0)

    May use tt.train_single_scheduler controlling the behavior more detail.

  • Model basically containing two parts:

    • Basic model is layers from input to embedding.
    • Model is Basic model + bottleneck layer, like softmax / arcface layer. For triplet training, Model == Basic model. For combined loss training, it may have multiple outputs.
  • Saving strategy

    • Model will save the latest one on every epoch end to local path ./checkpoints, name is specified by train.Train save_path.
    • basic_model will be saved monitoring on the last eval_paths evaluating bin item, and save the best only.
  • train.Train model parameters including basic_model / model. Combine them to initialize model from different sources. Sometimes may need custom_objects to load model.

    basic_model model Used for
    model structure None Scratch train
    basic model .h5 file None Continue training from a saved basic model
    None for 'embedding' layer or layer index of basic model output model .h5 file Continue training from last saved model
    None for 'embedding' layer or layer index of basic model output model structure Continue training from a modified model
    None None Reload model from "checkpoints/{save_path}"
  • Scheduler is a list of dicts, each containing a training plan. For detials please refer to keras_insightface

  • Restore training from break point

    from tensorflow import keras
    import losses, train
    data_path = 'datasets/faces_emore_112x112_folders'
    eval_paths = ['datasets/faces_emore/lfw.bin', 'datasets/faces_emore/cfp_fp.bin', 'datasets/faces_emore/agedb_30.bin']
    tt = train.Train(data_path, 'ghostnetv1_w1.3_s2.h5', eval_paths, model='./checkpoints/ghostnetv1_w1.3_s2.h5',
                    batch_size=128, random_status=0, lr_base=0.1, lr_decay=0.5, lr_decay_steps=45, lr_min=1e-5, eval_freq=1, output_weight_decay=1)
    
    sch = [
      # {"loss": losses.ArcfaceLoss(scale=32), "epoch": 1, "optimizer": optimizer},
      {"loss": losses.ArcfaceLoss(scale=64), "epoch": 35},
    ]
    tt.train(sch, initial_epoch=15)
  • We included all evaluation scripts in evaluation.ipynb

  • Evaluation example

    import evals
    basic_model = keras.models.load_model('checkpoints/ghostnetv1_w1.3_s2.h5', compile=False)
    ee = evals.eval_callback(basic_model, 'datasets/faces_emore/lfw.bin', batch_size=256, PCA_acc=True)
    ee.on_epoch_end(0)
    # >>>> lfw evaluation max accuracy: 0.996833, thresh: 0.223459, previous max accuracy: 0.000000, PCA accuray = 0.996000 ± 0.002494
    # >>>> Improved = 0.996833

    For training process, default evaluating strategy is on_epoch_end. Setting an eval_freq greater than 1 in train.Train will also add an on_batch_end evaluation.

    # Change evaluating strategy to `on_epoch_end`, as long as `on_batch_end` for every `1000` batch.
    tt = train.Train(data_path, 'ghostnetv1_w1.3_s2.h5', eval_paths, basic_model=basic_model, eval_freq=1000)

Other Basic Functions and Parameters

  • train.Train output_weight_decay controls L2 regularizer value added to output_layer.
    • 0 for None.
    • (0, 1) for specific value, actual added value will also divided by 2.
    • >= 1 will be value multiplied by L2 regularizer value in basic_model if added.
  • train.Train random_status controls data augmentation weights.
    • -1 will disable all augmentation.
    • 0 will apply random_flip_left_right only.
    • 1 will also apply random_brightness.
    • 2 will also apply random_contrast and random_saturation.
    • 3 will also apply random_crop.
    • >= 100 will apply RandAugment with magnitude = 5 * random_status / 100, so random_status=100 means using RandAugment with magnitude=5.
  • train.Train random_cutout_mask_area set ratio of randomly cutout image bottom 2/5 area, regarding as ignoring mask area.
  • train.Train partial_fc_split set a int number like 2 / 4, will build model and dataset with total classes split in partial_fc_split parts. Works also on a single GPU. Supports ArcFace loss family like ArcFace / AirFaceLoss / CosFaceLoss / MagFaceLoss.
  • models.buildin_models is mainly for adding output feature layer GDC to a backbone model. The first parameter stem_model can be:
    • String like ghostnetV1 / ghostnetV2 can be printed by models.print_buildin_models().
    • models.add_l2_regularizer_2_model will add l2_regularizer to dense / convolution layers, or set apply_to_batch_normal=True also to PReLU / BatchNormalization layers. The actual added l2 value is divided by 2.
    # Will add keras.regularizers.L2(5e-4) to `dense` / `convolution` layers.
    basic_model = models.add_l2_regularizer_2_model(basic_model, 1e-3, apply_to_batch_normal=False)
  • Gently stop is a callback to stop training gently. Input an n and <Enter> anytime during training, will set training stop on that epoch ends.
  • My history
    • This is a callback collecting training loss, accuracy and evaluating accuracy.
    • On every epoch end, backup to the path save_path defined in train.Train with suffix _hist.json.
    • Reload when initializing, if the backup <save_path>_hist.json file exists.
    • The saved _hist.json can be used for plotting using plot.py.
  • eval_folder.py is used for test evaluating accuracy on custom test dataset:
    python ./eval_folder.py -d {DATA_PATH} -m {BASIC_MODEL.h5}
    Or create own test bin file which can be used in train.Train eval_paths:
    python ./eval_folder.py -d {DATA_PATH} -m {BASIC_MODEL.h5} -B {BIN_FILE.bin}

Learning rate

  • train.Train parameters lr_base / lr_decay / lr_decay_steps / lr_warmup_steps set different decay strategies and their parameters.

  • tt.lr_scheduler can also be used to set learning rate scheduler directly.

    tt = train.Train(...)
    import myCallbacks
    tt.lr_scheduler = myCallbacks.CosineLrSchedulerEpoch(lr_base=1e-3, first_restart_step=16, warmup_steps=3)
  • lr_decay_steps controls different decay types.

    • Default is Exponential decay with lr_base=0.001, lr_decay=0.05.
    • For CosineLrScheduler, steps_per_epoch is set after dataset been inited.
    • For CosineLrScheduler, default value of cooldown_steps=1, means will train 1 epoch using lr_min before each restart.
    lr_decay_steps decay type mean of lr_decay_steps mean of lr_decay
    <= 1 Exponential decay decay_rate
    > 1 Cosine decay, will multiply with steps_per_epoch first_restart_step, epoch m_mul
    list Constant decay lr_decay_steps decay_rate
    # lr_decay_steps == 0, Exponential
    tt = train.Train(..., lr_base=0.001, lr_decay=0.05, ...)
    # 1 < lr_decay_steps, Cosine decay, first_restart_step = lr_decay_steps * steps_per_epoch
    # restart on epoch [16 * 1 + 1, 16 * 3 + 2, 16 * 7 + 3] == [17, 50, 115]
    tt = train.Train(..., lr_base=0.001, lr_decay=0.5, lr_decay_steps=16, lr_min=1e-7, ...)
    # 1 < lr_decay_steps, lr_min == lr_base * lr_decay, Cosine decay, no restart
    tt = train.Train(..., lr_base=0.001, lr_decay=1e-4, lr_decay_steps=24, lr_min=1e-7, ...)
    # lr_decay_steps is a list, Constant
    tt = train.Train(..., lr_base=0.1, lr_decay=0.1, lr_decay_steps=[3, 5, 7, 16, 20, 24], ...)

Mixed precision float16

  • Tensorflow Guide - Mixed precision
  • Enable Mixed precision at the beginning of all functional code by
    keras.mixed_precision.set_global_policy("mixed_float16")
  • In most training case, it will have a ~2x speedup and less GPU memory consumption.

Optimizers

  • SGDW / AdamW tensorflow_addons AdamW.
    # !pip install tensorflow-addons
    !pip install tfa-nightly
    
    import tensorflow_addons as tfa
    optimizer = tfa.optimizers.SGDW(learning_rate=0.1, weight_decay=5e-4, momentum=0.9)
    optimizer = tfa.optimizers.AdamW(learning_rate=0.001, weight_decay=5e-5)
    weight_decay and learning_rate should share the same decay strategy. A callback OptimizerWeightDecay will set weight_decay according to learning_rate.
    opt = tfa.optimizers.AdamW(weight_decay=5e-5)
    sch = [{"loss": keras.losses.CategoricalCrossentropy(label_smoothing=0.1), "centerloss": True, "epoch": 60, "optimizer": opt}]
    # Rectified Adam,a.k.a. RAdam, [ON THE VARIANCE OF THE ADAPTIVE LEARNING RATE AND BEYOND](https://arxiv.org/pdf/1908.03265.pdf)
    optimizer = tfa.optimizers.RectifiedAdam()
    # SGD with Lookahead [Lookahead Optimizer: k steps forward, 1 step back](https://arxiv.org/pdf/1907.08610.pdf)
    optmizer = tfa.optimizers.Lookahead(keras.optimizers.SGD(0.1))
    # Ranger [Gradient Centralization: A New Optimization Technique for Deep Neural Networks](https://arxiv.org/pdf/2004.01461.pdf)
    optmizer = tfa.optimizers.Lookahead(tfa.optimizers.RectifiedAdam())

Subcenter ArcFace

Evaluating on IJB datasets

  • IJB_evals.py evaluates model accuracy using insightface/evaluation/IJB/ datasets.
  • If any error occured from using the below scripts, kindly refer to evaluation.ipynb. Note: We can run the models on Google Colaboratory due to dependency issues.
  • In case placing IJB dataset IJB/IJB_release, basic usage will be:
    # Test mxnet model, default scenario N0D1F1
    python IJB_evals.py -m 'checkpoints/GhostFaceNet_W1.3_S1_ArcFace.h5' -d IJB/IJB_release -L
    
    # Test keras h5 model, default scenario N0D1F1
    python IJB_evals.py -m 'checkpoints/GhostFaceNet_W1.3_S1_ArcFace.h5' -d IJB/IJB_release -L
    
    # `-B` to run all 8 tests N{0,1}D{0,1}F{0,1}
    python IJB_evals.py -m 'checkpoints/GhostFaceNet_W1.3_S1_ArcFace.h5' -d IJB/IJB_release -B -L
    
    # `-N` to run 1N test
    python IJB_evals.py -m 'checkpoints/GhostFaceNet_W1.3_S1_ArcFace.h5' -d IJB/IJB_release -N -L
    
    # `-E` to save embeddings data
    python IJB_evals.py -m 'checkpoints/GhostFaceNet_W1.3_S1_ArcFace.h5' -d IJB/IJB_release -E
    # Then can be restored for other tests, add `-E` to save again
    python IJB_evals.py -R IJB_result/checkpoints/GhostFaceNet_W1.3_S1_ArcFace_IJBB.npz -d IJB/IJB_release -B
    
    # Plot result only, this needs the `label` data, which can be saved using `-L` parameter.
    # Or should provide the label txt file.
    python IJB_evals.py --plot_only IJB/IJB_release/IJBB/result/*100*.npy IJB/IJB_release/IJBB/meta/ijbb_template_pair_label.txt
  • See -h for detail usage.
    python IJB_evals.py -h

Evaluating on MegaFace datasets

  • Download megaface testsuite from baiducloud(code:0n6w) or gdrive. The official devkit is also included.
  • Prepare the environment. OpenCV 2.4 is required by the official devkit, for convenience, you can download it from BaiduCloud or GoogleDrive and unzip to /usr/local/lib/opencv2.4.
  • Edit and call run.sh to evaluate your face recognition model performance.

If errors

  • Run gen_megaface.ipynb or gen_megaface.py After making appropriate modifications if needed, i.e., provide model path; or move the GhostFaceNets directory to be a part of the MegaFace testsuite
  • Run run_remove_noises.sh
  • Run run_megaface.sh
  • Run run_megaface_refined.sh

Acknowledgement

This project includes code and ideas from the following sources:

We would like to thank leondgarse for their valuable contributions to the field and for sharing their work with the community. Their ideas and code have been instrumental in the development of this project and we are grateful for the opportunity to build upon their work.


Citation

If you use GhostFaceNets (or any part of this code in your research), please cite the following:

@ARTICLE{10098610,
  author={Alansari, Mohamad and Hay, Oussama Abdul and Javed, Sajid and Shoufan, Abdulhadi and Zweiri, Yahya and Werghi, Naoufel},
  journal={IEEE Access}, 
  title={GhostFaceNets: Lightweight Face Recognition Model From Cheap Operations}, 
  year={2023},
  volume={11},
  number={},
  pages={35429-35446},
  doi={10.1109/ACCESS.2023.3266068}}

and

  • BibTeX
    @misc{leondgarse,
      author = {Leondgarse},
      title = {Keras Insightface},
      year = {2022},
      publisher = {GitHub},
      journal = {GitHub repository},
      doi = {10.5281/zenodo.6506949},
      howpublished = {\url{https://github.com/leondgarse/Keras_insightface}}
    }
  • Latest DOI: DOI

  • BibTeX
    @misc{leondgarse,
      author = {Leondgarse},
      title = {Keras CV Attention Models},
      year = {2022},
      publisher = {GitHub},
      journal = {GitHub repository},
      doi = {10.5281/zenodo.6506947},
      howpublished = {\url{https://github.com/leondgarse/keras_cv_attention_models}}
    }
  • Latest DOI: DOI

About

This repository contains the official implementation of GhostFaceNets, State-Of-The-Art lightweight face recognition models.

https://ieeexplore.ieee.org/document/10098610

License:MIT License


Languages

Language:Python 69.2%Language:Jupyter Notebook 30.3%Language:Shell 0.5%