Metrics
Most knowledge distillation research primarily focuses on one metric: accuracy. Therefore, we have selected four leading knowledge distillation approaches from the previous research and evaluated them using various metrics, including accuracy, as well as F1 score, precision, and recall. Additionally, we are introducing a new metric for bias reduction through adversarial training.
Base Metrics
The majority of existing studies in knowledge distillation prioritize accuracy as the primary measure for evaluating performance. To offer a broader perspective on model performance, this research includes a variety of additional metrics in its assessment. The base metrics we evaluated encompass top-1 accuracy, precision, recall, and the f1 score.

Debiasing The Network Using Adversarial Feature Attacks
To Reduce Bias, We Incorporate An Adversarial Attack
The traditional Knowledge Distillation (KD) framework, which consists of a teacher-student model, is augmented with an adversary model to effectively minimize bias. The student model is primarily focused on the task of image classification prediction. In parallel, the adversary model is designed to detect biases by predicting sensitive attributes from the student model’s outputs, with a particular emphasis on gender in this case. The adversary model aims to refine the student model to the extent to which the adversary can no longer accurately predict these sensitive attributes, thereby reducing bias.
​
To regulate the influence of this adversarial method, a hyperparameter known as lambda (λ) is introduced. λ acts as a debiasing weight, determining the strength of the adversarial influence on the student model. During the training process, both the student and adversary models are simultaneously developed. The student model is refined to minimize the divergence and cross-entropy (CE) loss. Concurrently, the adversarial loss is maximized, thereby minimizing the adversaries ability to predict the sensitive attributes of the student model output. As a result, this methodology not only optimizes the student model for accuracy but also ensures fairness, leading to a more balanced and unbiased machine learning model.
