Introduction
Traditional methods of human identification relying on credentials (e.g., PINs, identification documents, etc) are unable to meet the growing demands for stringent security in applications such as access control, surveillance, and identification. Alternatively, biometrics (e.g., Fingerprint, Iris, Vein, Voice, Face, etc) gained significant attention over the past two decades as an alternative solution. Face verification (FV) as a biometric authentication method is particularly significant. It offers several advantages over other biometric traits. For instance, it is non-intrusive and easy to use due to the widespread availability of cameras in a variety of devices and environments. It is widely used in several applications, including security systems and mobile devices, to provide secure access to personal devices and sensitive information.
Discriminative feature extraction from image/video data is of paramount importance in face verification. Early studies [1–4] focused on utilizing hand-crafted feature extraction methods. Later, deep learning based methods revolutionized feature extraction by autonomously learning intricate patterns and representations from data automatically [5–7]. In particular, network architecture and loss function play crucial and complementary roles in feature extraction. The architecture defines the network’s structure and capacity to learn representations, while the loss function guides the learning process towards extracting meaningful and discriminative features, relevant to the task at hand. The synergy between these components is essential for achieving optimal performance.
From a loss function design perspective, SphereFace [8] introduces a multiplicative angular margin loss function. CosFace [9] demonstrates the effectiveness of subtracting a margin rather than multiplying it. ArcFace [10] proposes an additive margin approach, which facilitates the learning of discriminative features. However, the limitation of SphereFace [8] is that its performance is sensitive to hyperparameters, particularly the choice of the angular margin parameter. Indeed, improper tuning of this parameter may lead to suboptimal feature discrimination. CosFace [9] exhibits computational complexity, especially when dealing with large-scale datasets. Specifically, the computation of the cosine similarity with respect to each class centroid during training can become computationally expensive, potentially affecting training times and resource efficiency. ArcFace [10] relies on hyperparameters that necessitate manual selection, involving repeated trial and error, which is time-consuming. Thus, on one hand, the hyperparameters of angular margin loss functions offer leverage for tuning deep learning architectures to achieve improved results. On the other hand, improper selection of these hyperparameters may substantially impact accuracy.
To the best of our knowledge, to date, AdaptiveFace [11] and AdaCos [12] are the primary methods offering automated hyperparameter selection for large margin loss frameworks. AdaptiveFace [11] estimates the margin hyperparameter, while AdaCos [12] focuses on the scale hyperparameter. Several related approaches address various aspects of hyperparameter optimization. For example, HAMFace [13] introduces a Hardness Adaptive Margin Loss, applying larger margins to challenging positive samples and using a control hyperparameter to jointly manage the margin and scale hyperparameters. The study by Gim et al., [14] incorporates Noise Direction Regularization (NDR), leveraging the L2 norm of the features as a regularizer and introducing a hyperparameter to improve performance. CoReFace [15] further advances the field by employing an adaptive margin-based contrastive loss, which utilizes an exponential moving average for the margin, while relying on manual scale tuning. To overcome the limitations of existing methods, this study proposes AML, a novel framework that automates the selection of both margin and scale hyperparameters. This estimates the margin using input dissimilarity information derived from the cross-product of the feature and weight vectors, while the scale hyperparameter is determined based on the number of classes and the effective range of AML (more details in [sec:AML]Adaptive margin loss).
From a network architecture design perspective, the Deep Residual Network (ResNet) [16] introduced skip connections, where each micro-block (residual function) includes a residual path. This path element-wise adds input features to the output of the same micro-block, defining it as a residual unit. ResNet evolved into various architectures such as Inception-Resnet [17] and ResNeXt [18], each characterized by changes in their structural designs. Later, [19] proposed the Dense Convolutional Network (DenseNet), which uses dense connections, with each layer receiving inputs from all preceding layers and passing its own feature maps to all subsequent layers, thus enhancing information flow and feature reuse. Chen et al., [20] explore the relationship between residual networks and densely connected networks, demonstrating that residual networks exhibit characteristics of densely connected networks when connections are shared across layers. They proposed the Dual Path Network (DPN), which leverages the residual path for implicit feature reuse and continuous exploration of new features through the densely connected path.
Several studies [16–20] rely on downsampling in the network architecture, however, this often leads to significant information loss. García et al., [21] highlighted the challenges of preserving diagnostic features in medical images, proposing feature-preserving techniques, which, despite improvements, were unable to completely prevent degradation. Similarly, Sedlar et al., [22] noted that inappropriate downsampling in genomic signal processing obscured critical patterns, advocating for tailored rules to optimize the balance between data reduction and feature retention. Zhao et al., [23] introduced random shifting to minimize systematic information loss in CNN downsampling layers, enhancing feature detail preservation. Kuen et al., [24] proposed stochastic downsampling to improve network regularization and enable cost-adjustable inference, however, this introduces variability, which could affect feature consistency. To address this limitation, we introduce the Dual Path Network+ with a carefully redesigned Block-1, which includes a series of 3x3 convolutions, batch normalization, and ReLU activation functions to enhance spatial resolution and reduce computational costs. Moreover, DPN+ maintains shared connections across layers to maximize the utilization of discriminative features, thereby enhancing both performance and efficacy. More details can be found in [sec:IDPN]Dual path network+.
In summary, existing approaches suffer due to the requirement for manual hyperparameter selection, which significantly impacts on their performance, or alternatively, facing computational challenges, when it comes to dealing with large-scale, real-world datasets. Moreover, due to the use of downsampling in the CNN architectures, there is significant information loss, reducing the discriminating ability of the underlying features. Fig 1 provides an overview of the proposed AMD-FV framework, incorporating Dual Path Network+ and Adaptive Margin Loss. This achieves state-of-the-art performance in face verification tasks across five diverse datasets, varying in volume and complexity.
[Figure omitted. See PDF.]
Adaptive Margin Loss is used in adaptive estimation of the margin and scale hyperparameters. The Dual Path Network+ offers increased spatial resolution at reduced computational costs, via shared connections across layers to maximize the utilization of discriminative features. Data preprocessing includes facial landmarks detection, face alignment using five key facial points (i.e., nose, eyes, and corners of the lips), and cropping. Next, the cropped image undergoes data augmentation to provide the network with various versions of the same input, facilitating the learning of discriminative features.
The contributions of this work are as follows:
* Adaptive estimation of the margin hyperparameter in large margin loss is achieved. The proposed method utilizes the dissimilarity of input data, estimated via the cross product of the feature and weight vectors, to adaptively adjust the marginalization of features.
* The scale hyperparameter of the large margin loss is automatically selected by leveraging the number of the classes and AML range. This adaptive approach scales the marginalized features for improved classification, enhancing the effectiveness of the model.
* The Dual Path Network+ is introduced with a redesigned Block-1, which enhances spatial resolution and reduces computational costs. This block incorporates a series of 3x3 convolutions, batch normalization, and ReLU activation functions. Moreover, the architecture leverages shared connections across layers to maximize the utilization of discriminative features, ultimately improving both performance and efficacy.
The remainder of the paper is organized as follows. In Section Related works, related works in the literature are discussed and analyzed. Section Materials and methods describes the details of the proposed method. Section [sec:exp]Experiments illustrates the experimental setup, presents the details of the utilized datasets and reports on the implementation details. Section [sec:results]Results presents a thorough ablation study, followed by a performance comparison with related works, and a discussion of the key findings. Finally, conclusions and future work are drawn in Section [sec:conclusion]Conclusions.
Related works
Margin-based loss functions, such as SphereFace [8], CosFace [9], and ArcFace [10], demonstrated significant success in face verification tasks. The hyperparameters associated with these loss functions, often manually chosen, play a crucial role on the performance of the deep learning model. Indeed, suboptimal or incorrect selection of these hyperparameters can significantly affect their properties and, subsequently, model performance. Typically, determining these parameters involves a time-consuming and laborious trial-and-error process, the success of which is limited by the choice of initial parameters. To address this issue, researchers have proposed data-driven methods to estimate these hyperparameters. For instance, Liu et al., [11] argue that face data typically consists of large-scale classes and samples, emphasizing that selecting appropriate classes and sample numbers for training helps the model achieve better generalization. They proposed the use of hard prototype mining and introduced the AdaptiveFace model to estimate the associated hyperparameters. However, the computational overhead of learning individual margins for each class thoroughly considered. In the case of large-scale datasets, this could potentially lead to significant increases in training time and memory requirements. Moreover, as demonstrated in the experiments, the performance of the method is sensitive to the choice of the hyperparameter controlling the strength of the margin constraint. Zhang et al., [12] propose the Adacos model to tackle the challenge of hyperparameter selection in face verification tasks. However, their approach presents a number of challenges. First, it assumed that during training, the sum of logits for non-corresponding classes remains almost constant. Beyond empirical evidence, the theoretical justification for this assumption, crucial in deriving the adaptive scale parameter, is not proven. Moreover, the analysis of probability curves is based on simplified models, which may not fully capture the complexity of real-world face recognition scenarios, e.g., class imbalance and samples of varying difficulty.
Zhang et al., [25], propose diversity regularization by training the margin hyperparameter and constraining it within the range of 0.5 and . While the authors automatically train the margin hyperparameter, they introduce two additional manual parameters in terms of the range. Moreover, there is insufficient analysis of the impact of the balance hyperparameter across different datasets and network architectures, and the computational overhead associated with the calculation of adaptive margins and diversity regularization. RamFace [26] addresses racial bias by proposing an adaptive margin loss based on intra-identity compactness using multitask learning. Two loss functions, i.e., race adaptive margin-based loss and race classification loss, are introduced, however, the additional hyperparameter is used to balance them. There is insufficient consideration on the appropriate choice of this hyperparameter, and its impact on model performance. In addition, the manuscript neither provides formal guarantees on the fairness properties of the proposed method, nor does it discuss in sufficient detail the potential trade-offs between overall accuracy and fairness across different racial groups. EnhanceFace [27], proposes adaptive weighted softmax loss by segregating hard samples into semi-hard and harder samples. The loss function adjusts the corresponding weights during training. The authors introduce the enhance rate , used to adjust the effects of semi-hard samples. However, the selection of this parameter appears arbitrary, as there is lack of justification for the chosen range and the step size. JAMsFace [28] proposes a joint adaptive margin loss function, which learns class-related margins by incorporating margin penalties within the cosine angle of a loss function. These models mitigate the problem to some extent. However, they only provide a means of estimating a single hyperparameter, i.e., margin, and moreover, there is a lack of a clear formulation on how this is calculated based on class distribution. In contrast, the proposed AML approach estimates both the scale and margin hyperparameters.
In the context of CNN architectures, García et al., [21] discussed the difficulty of preserving diagnostic features in medical images during downsampling. He et al., [29] introduced the Residual Network (ResNet), which utilizes skip connections to enable very deep networks, addressing optimization challenges. However, the use of downsampling in ResNet leads to significant information loss. Sedlar et al., [22] identified similar challenges in genomic signal processing, calling for strategies that balance data reduction with feature retention. Zhao et al., [23] proposed random shifting to minimize systematic information loss during downsampling. In contrast, Huang et al., [30] introduced the Dense Convolutional Network (DenseNet), which uses skip connections to concatenate features from earlier layers, but at the expense of high computational complexity. Chen et al., [20] demonstrated that ResNet reuses features via residual paths, while DenseNet explores new features through densely connected paths. Building on this, Chen et al., [20] proposed the Dual Path Network, which combines the strengths of both ResNet and DenseNet, although it still suffers from information loss due to its downsampling approach. The proposed Dual Path Network+ is an enhancement to the original DPN, improving spatial resolution and computational efficiency, while leveraging shared connections across layers to maximize feature utilization.
Materials and methods
Background
Angular margin loss function extends Cross-Entropy Softmax Loss (CESL) through the introduction of hyperparameters, which is given as follows:
(1)
where xi denotes the embedded feature vector of the sample of the yi class, are the and columns of the weights, respectively, whereas are the respective biases. N and n denote the batch size and the number of classes, respectively.
The CosFace model [9] proposes a Large Margin Cosine Loss (LMCS) and removes the radial variation of the features and weights [9]. The last fully connected (FC) layer operates on the normalized features xi and the normalized weights Wj. The norm of is set to 1 and the norm of is replaced by the hyperparameter s. Hence the activation of the last FC layer is given by . LMCL is given by:
(2)
where s and denote the scale and margin hyperparameters respectively.
Adaptive margin loss
In this section, we introduce the proposed angular margin-based loss function, AML, which streamlines the selection process for the scale hyperparameter s and margin hyperparameter in the CosFace loss function. The inspiration for this approach stems from the observation that humans naturally take into account dissimilarities when comparing images. For instance, when presented with two images, e.g., one of a male and the other of a female, individuals instinctively focus on their differences rather than similarities to distinguish between them. This observation suggests that the concept of dissimilarity could serve as a basis for matching. In contrast, traditional neural networks typically measure the degree of similarity between the presented input patterns and the network weights.
In this research, we leverage the concept of dissimilarity akin to human visual perception to introduce a novel methodology for the automated estimation of the margin hyperparameter. The dissimilarity of a batch of images is assessed based on the magnitude of the cross product, subsequently subtracting it from the feature map. While the proposed approach adopts the architecture of CosFace, it differs significantly, as CosFace relies on a manually selected margin value. For the automated estimation of the scale hyperparameter, we consider the product of the number of classes within the training dataset and the range of the loss function. The proposed AML loss function operates on the normalized features xi and the normalized weights Wj of the last FC layer. It automatically estimates both the margin and the scale hyperparameters.
For the adaptive estimation of margin hyperparameter m, we estimate the mean dissimilarity of the feature vectors xi and weight vectors in the given batch of images by using the magnitude of the cross product (see A. Magnitude of Cross Product for more details). We consider the mean distance as the probable average dissimilarity of the batch n and estimate the margin hyperparameter m as follows:
(3)
For the adaptive estimation of the scale hyperparameter s, the logarithm of the number of classes and the maximum range is used as follows:
(4)
where N is the number of classes, and R is the maximum range of AML (see B. Range of AML for more details). The above choice is motivated by [31], where it is argued that the logarithmic function is effective in scaling down large values, thus facilitating easier mathematical operations. Conversely, the exponential function is utilized to provide leverage, enabling the estimation of the rate of change with precision within small increments. In our scenario, we deal with a large dataset (MS1Mv2) comprising 85,000 classes, with the activation of the last fully connected layer being in the range of 0 to 1. Classifying 85,000 classes within the spectrum of [0,1] presents a considerable challenge. Therefore, the proposed method works by appropriately scaling the marginalized features using the logarithmic function to mitigate the impact of the large number of classes. Subsequently, we utilize the exponential function to emphasize the small range (i.e., 0 to 1) associated with the activation of the last FC layer. The steps of the proposed AML are outlined in Algorithm 1. The Adaptive Margin Loss function LAML with the automatically selected margin hyperparameter m and scale hyperparameter s is as follows:
(5)
Algorithm 1. Adaptive Margin Loss Algorithm.
Require: Feature map xi and weights Wj of the last fully
connected layer.
Ensure: The updated loss via the estimated scaled and
marginalized parameters.
1: Normalize the feature map xi in the batch.
2: Normalize the weights Wj in the batch.
3: Update the last FC layer through the normalized feature map
and the normalized weights.
4: Estimate the margin hyperparameter value via Eq 3.
5: Estimate the scale hyperparameter value via Eq 4, where N
is the number of classes, and R is the maximum range of AML
(see B. Range of AML for details).
6: Update the output of the FC layer by multiplying the scale
hyperparameter and subtracting the margin hyperparameter.
7: Update the loss using Eq 5.
Dual path network+
As discussed in [24, 22, 21], and [32], downsampling in CNN represents a trade-off between low-level spatial information and richer high-level semantic information. Therefore, in the proposed method, to prevent information loss, we refrain from downsampling and instead, leverage the low-level features in the first convolutional stage of DPN. This approach leads to the development of the proposed Dual Path Network+, which is a 68-layer convolutional neural network inspired by [20]. The overall neural network structure of DPN+ is presented in Table 1.
[Figure omitted. See PDF.]
DPN+ is constructed by stacking five stages of convolution, as outlined in Table 1. The structure of the first block comprises a series of three convolutional layers, where each layer consists of a 3 3 convolution with a stride of 1, followed by Batch Normalization [33] and ReLU activation [34]. This initial stage leverages low-level feature information as the foundation for generating highly discriminative features. In the first block of DPN+, information loss due to downsampling is mitigated, and low-level feature information is enhanced through a series of convolutional layers.
The subsequent stages of the network adhere to the DPN neural network structure, employing 32 convolutional layers in each sub-block, similar to ResNeXt [35]. The structure of each sub-block follows a bottleneck style [29], comprising a convolutional layer followed by a convolutional layer, and concluding with another convolutional layer. The final convolutional layer is divided into two paths, one connecting the residual path and the other linking the densely connected path.
Experiments
In this section, we present the details of the datasets used in the development and experimental verification of the proposed AMD-FV. This is followed by a brief overview of the system implementation.
Datasets
Extensive experiments on five well-known FV benchmarks were conducted, including Labeled Faces in the Wild (LFW) [36], Megaface [37], IARPA Janus Benchmark B (IJB-B) [38], Cross-Age LFW (CALFW) [39], and Cross-Pose LFW (CPLFW) [40]. These datasets encompass diverse conditions such as pose, gender, occlusion, and illumination variations, detailed statistics of the datasets are provided in Table 2.
[Figure omitted. See PDF.]
Implementation details
This section systematically provides the implementation details, including the neural network training configuration, data preprocessing pipeline, evaluation protocols, and the computational environment. The code is available at https://github.com/zeeshanahmedkhan/AML.
Training Configuration: The experiments were conducted on a hardware setup of three NVIDIA TITAN X GPUs interconnected via NVLink, with a global batch size of 128 (distributed as 42-43 samples per GPU). Synchronized batch normalization was used to ensure consistent gradient updates across GPUs. The learning rate began at 0.1 with a 5-epoch linear warmup phase, followed by a cosine annealing schedule that gradually reduced the rate to a minimum of 1e-6. Stochastic Gradient Descent (SGD) with momentum ( = 0.9) was employed as the optimizer, coupled with gradient clipping to a maximum L2 norm of 5.0 to prevent exploding gradients. The regularization strategy included weight decay (5e-4 applied exclusively to the convolutional and fully connected layers), label smoothing ( = 0.1), and stochastic depth with a DropPath probability of 0.2.
Data Preprocessing Pipeline: Face detection was performed using MTCNN [41] with 5-landmark localization (eyes, nose, and mouth corners). Alignment involved a similarity transformation based on eye coordinates, supplemented by 3D face reconstruction for images with extreme poses (yaw or pitch exceeding 45∘). Augmentation strategies included random horizontal flipping (50% probability), color jitter (brightness 0.2, contrast 0.15), and patch masking (up to 15% of the image area). Input normalization comprised channel conversion, per-channel mean subtraction ([91.4953, 103.8827, 131.0912]), and 8-bit quantization with histogram equalization.
Evaluation Protocols: For standard benchmarks such as LFW [36], CALFW [39], and CPLFW [40], we adopted a 10-fold cross-validation protocol with unrestricted settings. MegaFace [37] evaluation followed the “Large” protocol (1 million distractor images), while IJB-B [38] verification metrics (TAR@FAR) were computed over 10,000 genuine and 8 million impostor pairs. All face embeddings were compared using cosine similarity, with decision thresholds optimized via the ROC convex hull method.
Computational Environment: The implementation used PyTorch 1.12.1 with CUDA 11.6, leveraging mixed FP16/FP32 precision through NVIDIA Apex AMP. Parallelization was managed via PyTorch’s DataParallel module with NCCL backend, and all random processes were fixed to a seed value of 42 for reproducibility purposes.
Results
In this section, we provide an extensive ablation study, followed by a performance comparison with state-of-the-art models for face verification. The ablation study demonstrates the impact of the proposed DPN+ architecture and AML on system performance across the various datasets. Last, we provide a discussion of the main outcomes of performance evaluation.
The performance metrics used in the subsequent sub-sections are as follows:
True Positive Rate (TPR): TPR is defined as the ratio of samples identified as true positives (TP) over the total number of positives (P):
The total number of samples identified as positive is given by:
where FN relates to the false negatives.
False Positive Rate (FPR): FPR is defined as the ratio of samples identified as false positives (FP) over the total number of positives (P):
Mean Verification Accuracy (Accuracy): This metric measures the overall correctness of the model in verifying identities. It is defined as the ratio of correctly predicted instances to the total instances:
Precision: Precision quantifies the accuracy of the positive predictions made by the model. It is calculated as the ratio of true positives to the sum of true positives and false positives:
Recall: Recall measures the model’s ability to identify all relevant instances. It is defined as the ratio of true positives to the sum of true positives and false negatives:
True Acceptance Rate (TAR): TAR is measured as the ratio of the number of positive matches, Np, over the sum of the number of positive matches and the number of false rejects, Nr:
False Acceptance Rate (FAR): A complimentary metric to TAR, FAR relates to the proportion of false accepts, i.e., FP, which are incorrectly accepted as genuine attempts, over the total number of impostor attempts, i.e., the sum of FP and TN:
Rank-1 score: This metric provides a measure of the FV system’s ability to correctly identify the top match for a given query image from a set of images. It is defined as the proportion of queries where the correct match appears as the top-ranked candidate in the system’s output:
where M and Q relate to the number of queries where the true match ranks as top, and the total number of queries, respectively.
Ablation study
The ablation study provides evidence in support of the justification for the key components of AMD-FV. Five models are evaluated, using the architecture of a 68 layered DPN+ neural network. Their summary descriptions are shown in Table 3.
[Figure omitted. See PDF.]
The models were analyzed according to several performance metrics, such as model complexity (i.e., parameter number and model size), computational time measured in FLOPs, inference time in seconds, and accuracy rate. Note that model complexity, computational time, and inference time are closely related to the computational power of the machine, whereas accuracy rate is related to model performance on the input data. Tables 4 and 5 illustrate the computational complexity and inference times of the evaluated models, respectively. It should be noted that the same CNN architecture was used in all of the models. Therefore, the chosen metrics will have similar values, providing the means to perform a fair comparison of impact of loss functions on the basis of the accuracy rate.
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
Robustness of DPN+
In this sub-section, we focus on the analysis of the proposed Dual Path Network+ over the Dual Path Network, so as to provide a fair comparison. Both DPN and DPN+ were trained on the same dataset (MS1Mv2 [10]), under the same settings (See Section [sec:Implementation]Implementation details for details) and loss function (i.e., CosFace loss [9] ). The models were then evaluated on the five benchmark face datasets. The obtained results support the claim that the proposed DPN+ model (DPN+68-C) extracts more discriminative features than those of the DPN model (DPN68-E). The verification accuracy of DPN+68-C is 0.17%, 1.78% and 2.94% higher than that of DPN68-E on the LFW, CPLFW and Megaface datasets, respectively ([Table_DPN_vs_DPN+]Table 9). A slight drop of 0.07% in the performance of the DPN+68-C compared to DPN68-E is observed in the case of the CALFW dataset. The slight drop in performance of the DPN+68-C model in [Table_DPN_vs_DPN+]Table 9 on the CALFW dataset occurs because the model has not been specifically trained to handle age gaps in photos of the same person. While it excels at matching faces with varying angles (CPLFW), crowded scenes (MegaFace), and general recognition (LFW), aging variations such as wrinkles, facial shape-shifts are distinct and unique challenges. To address this, state-of-the-art approaches employ dedicated techniques, e.g., synthetically "aged" face examples during training, which the proposed model does not provide for. However, DPN+’s stronger performance on pose variations and complex real-world scenarios suggests it is better suited for practical applications, where camera angles and crowded environments matter more than extreme age differences. To address aging variations within the current framework, it is possible to augment the training set with age-specific instances or perform fine-tuning for aging effects, without compromising on the model’s current strengths. Finally, the ROC curves in [fig:ROC_DPN_vs_DPN+]Fig. 2 demonstrate that the DPN+68-C model significantly surpasses the DPN68-E model on the IJB-B dataset.
[Figure omitted. See PDF.]
Robustness of AMD-FV
Next, we provide a comparison between the most relevant state-of-the-art angular loss function models, and the proposed AMD-FV model. The SphereFace model [8] proposes a multiplicative angular margin, where the margin value is multiplied by the angle, which, in turn, provides for a fixed distance between the samples. The multiplicative property of the margin highly influences the angle. When the angle decreases towards zero, the margin vanishes (see Fig. 5(a)). The CosFace[9] and ArcFace models [10] introduce the additive angular margin. In CosFace model [9], the margin value is subtracted from the cosine value (Eq. 2), which overcomes the limitation of the vanishing margin (see Fig. 5(b)). Compared with the CosFace model [9], the ArcFace model [10] adds a margin to the angle. The ArcFace model [10] also has a vanishing margin problem. However, because of the additive property, a fixed margin value does not significantly affect the angle (see Fig. 5(c)). In the proposed model, AML automatically estimates the margin using the input data, so that it does not vanish. Specifically, it uses the cross product of the feature map to estimate the margin according to Eq. 3 and subtracts the margin from the cosine value (Eq. 5) (see Fig. 5(d)). By adaptive estimation of the margin and the scale hyperparameters, AML provides an alternative to the otherwise, time-consuming training of the trial and error approach.
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
(a) SphereFace [8], (b) CosFace [9], (c) ArcFace [10], and (d) AMD-FV. Please note that the setting of binary classification is considered.
The principal objective in this research is the development of an automated procedure for the selection of the margin and scale hyperparameters, which resulted to the proposed AML function. Analysis of the experimental results demonstrates that the accuracy obtained by the AMD-FV model surpasses the performance of both state-of-the-art AdaptiveFace [11] and Adacos models [12]. Note that the mean verification accuracy on the LFW dataset is almost saturated and thus, a small increase in the accuracy has a significant impact. The verification accuracy achieved by the AMD-FV model is 99.75%, which is higher than that of AdaptiveFace at 99.63% and Adacos at 99.71% ([Table_AML-DPN+]Table 6). In addition, the accuracy of the AMD-FV model is higher than that of the baseline model on all the testing datasets. It is worth noting that the introduction of AML results to an average improvement of 2.338% compared to using CESL ([Table_baseline_vs_AML-DPN+]Table 10 and [fig:ROC_AML-DPN+_vs_baseline]Fig. 3).
[Figure omitted. See PDF.]
The verification accuracies obtained by the DPN+68-B model and the DPN+68-D model are similar to one another (Table 8). Specifically, the use of ArcFace loss led to outperforming the DPN+68-D model by a margin of 0.19%, on average, suggesting that the proposed adaptive DPN+68-D model not only surpasses the state-of-the-art adaptive margin-based models, but also favorably compares with the manually optimized margin-based model (i.e., ArcFace). Fig. 4 provides a visual comparison of the computational complexity of the DPN+68-D model and the ArcFace model [10] (i.e., ResNet with 100 layers). It is shown that the proposed AMD-FV model leads to substantial reductions of 32%, 26.92% and 41.91% in terms of the number of layers, parameters and FLOPS, respectively. Moreover, Table 7 compares the performance of two DPN+ architectures, i.e., DPN+68-B (trained with ArcFace loss) and DPN+68-D (trained with AML loss). The results reveal a marginal performance differential between the two models. DPN+68-D exhibits near-equivalent accuracy to DPN+68-B, with minor deviations, i.e., -0.03% on LFW, -0.19% on CALFW, and -0.35% on CPLFW. Notably, DPN+68-D outperforms DPN+68-B by +0.28% on the IJB-B benchmark.
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
Tables 7, 8, 11 and 13 demonstrate that ArcFace achieves marginally superior performance on certain benchmarks, however, this apparent advantage stems from ArcFace’s reliance on manually optimized hyperparameters, performed on each of the employed datasets. In contrast, AML eliminates the need for manual tuning by automatically adapting the margin and scale parameters based on input dissimilarity and class distribution. This automation inherently introduces trade-offs as follows:
[Figure omitted. See PDF.]
Manual Tuning vs. Automation: ArcFace’s manual margin selection allows fine-grained optimization for benchmark metrics, however, it requires extensive trial-and-error, which is impractical for large-scale, real-world deployment. Instead, AML replaces this process with data-driven adaptation, sacrificing marginal accuracy gains for generalizability and scalability.
Computational Efficiency: As highlighted in Fig. 5, AML respectively reduces model complexity by 32%, 26.9% and 41.9% in terms of the number of layers, parameters, and FLOPs compared to ArcFace. This efficiency makes AML more suitable for resource-constrained environments.
Performance Contextualization: The slight performance gap is negligible in practical applications, however, it underscores the cost of automation. AML avoids the laborious hyperparameter search, making it a preferable choice for dynamic datasets, where manual tuning is infeasible.
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
Comparison with state-of-the-art methods
In this sub-section, we consider the performance of AMD-FV in regards to related state-of-the-art approaches. The mean verification accuracies acquired on the LFW dataset are presented in Table 11. The proposed AMD-FV method attains a competitive mean verification accuracy of , surpassing the performance of the majority of previously published methods e.g., [42, 8, 43], and [44]. Arcface [9] exhibits a slightly higher accuracy compared to AMD-FV by 0.08%. However, in comparison to the manually optimized CosFace model [9], the proposed AMD-FV method with automated hyperparameter selection demonstrates an improvement of 0.02%. This suggests that the automated hyperparameter selection process of AMD-FV could serve as an alternative to the empirically determined hyperparameter values of the CosFace model, eliminating the need for time-consuming trial and error. Furthermore, in comparison to the other two adaptive models, namely AdaptiveFace [11] and Adacos [12], AMD-FV demonstrates a marginal improvement of 0.13% and 0.04%, respectively.
IJB-B is considered to be a challenging dataset concerning occlusion, poses, and illumination. In Table 12, the True Acceptance Rate (TAR) is presented concerning various False Acceptance Rates (FAR) on the IJB-B dataset. The proposed AMD-FV model outperforms all other models. Specifically, it improves on the VGGFace2 model by 6%, and by 4.6% and 3.9% compared to MN-V and MN-VC, respectively, at a FAR of 0.001. At a FAR of 0.01, the proposed AMD-FV surpasses VGGFace2, MN-V, and MN-VC models by 2.83%, 2.23%, and 1.93%, respectively. At a FAR of 0.1, AMD-FV’s performance marginally outperforms that of the VGGFace2, MN-V, and MN-VC. This insight underscores the improved performance of AMD-FV compared to state-of-the-art methods on the IJB-B dataset.
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
The Rank-1 identification scores of existing methods on the Megaface dataset are presented in Table 13. The gallery set of 1M distractor images in Megaface presents a wide range of poses, ages, and illumination changes, making this dataset exceptionally challenging for face verification. Notably, the proposed AMD-FV method outperforms several state-of-the-art models, such as the CosFace model [9], Sphereface model [8], and FaceNet [42], with considerable margins (i.e., greater than 10%). However, when compared to the adaptive hyperparameter estimation models of AdaptiveFace [11] and Adacos [12], the results of AMD-FV are lower. This difference is attributed to Megaface’s use of a large number of distractor images in its verification protocol, which affects AMD-FV’s margin hyperparameter estimation values computed based on the mean. It is worth noting that the primary objective of this work is to adaptively estimate the large margin cosine loss of CosFace [9]. In this regard, the proposed AMD-FV model achieves a Rank-1 score of 92.16%, significantly surpassing CosFace by 9.44%.
The mean verification accuracy on the Cross Age LFW (CALFW) [39] and Cross Pose LFW (CPLFW) datasets [40] is presented in Table 14. These datasets extend the LFW dataset by introducing variations in age and pose, thereby increasing the complexity level. On the CALFW dataset, AMD-FV significantly outperforms the Center loss [47], SphereFace [8], and VGGFace2 models [5] by 10%, 5.21%, and 4.94%, respectively. Similarly, on the CPLFW dataset, the mean verification accuracy of AMD-FV surpasses that of the Center loss [47], SphereFace [8], and VGGFace2 models [5] by 13.52%, 9.60%, and 7.00%, respectively.
[Figure omitted. See PDF.]
Discussion
A fair comparison was conducted to evaluate the performance of the proposed Dual Path Network+ over the Dual Path Network. Both DPN and DPN+ were trained using the same dataset (MS1Mv2 [10]), under identical settings (see [sec:Implementation]Implementation details for more details).
The proposed Adaptive Margin Loss automatically estimates the margin using the input data. Specifically, it utilizes the cross product of the feature map to estimate the margin according to Eq 3 and then subtracts the margin from the cosine value (Eq 5) (see Fig 5(d)). Through adaptive estimation of the margin and scale hyperparameters, AML offers an alternative to the conventional time-consuming trial-and-error training approach.
The overall results demonstrate that the proposed AMD-FV model outperforms the majority of the state-of-the-art methods. This advancement stems from three primary contributions:
DPN+ Contribution: [Table_DPN_vs_DPN+]Table 9 demonstrates that DPN+68-C outperforms DPN68-E by 0.17% on LFW and 1.78% on CPLFW. This improvement stems directly from DPN+’s redesigned Block-1, which preserves low-level spatial features through three sequential 3x3 convolutions (vs. DPN’s 3x3 maxpool with stride=2). The absence of downsampling in Block-1 reduces information loss, particularly benefiting pose-sensitive datasets like CPLFW.
AML Contribution: Table 7 reveals that AML in DPN+68-D model achieves comparable accuracy to ArcFace loss in DPN+68-B model despite the use of automated hyperparameter selection (i.e., 99.75% vs. 99.78%) on LFW. This marginal drop of 0.03% reflects the cost of replacing manual tuning with AML’s data-driven adaptation. However, AML reduces training complexity by eliminating tedious and laborious effort on manual hyperparameter search.
Integrating DPN+ with AML: The combined AMD-FV model achieves 41.9% lower FLOPs than ArcFace (Fig. 4) by combining AML’s efficient scaling with DPN+’s parameter reduction.
Conclusions
This work presents AMD-FV, an enhanced deep-learning model for face verification, which capitalizes on two key improvements. The first contribution addresses the issue of information loss caused by downsampling inherent in the Dual Path Network. We introduce the Dual Path Network+, incorporating a novel Block-1 with 3x3 convolutions, batch normalization, and the ReLU activation function. DPN+ offers a favorable alternative in terms of computational cost, including fewer layers, fewer parameters, and lower FLOPS than ResNet. The second contribution addresses the tedious and time-consuming process of trial and error when selecting the margin and scale hyperparameters of angular loss functions. We propose the Adaptive Margin Loss function, which automates hyperparameter selection, eliminating the need for extensive experimentation typically required by state-of-the-art methods. Overall, the AMD-FV model demonstrates competitive performance with the advantages of automated hyperparameter selection. Specifically, the proposed framework achieves verification accuracy of 99.75% on the LFW dataset, outperforming AdaptiveFace (99.62%) and Adacos (99.71%). Moreover, it delivers TAR of 94.8%, 97.3% and 98.9% at FAR of 0.001, 0.01 and 0.1, respectively, on the IJB-B dataset, surpassing VGGFace2 by up to 6%. Similarly, competitive performances are delivered on Megaface with Rank-1 identification scores of 92.16%, compared to CosFace, and MVA of 95.51% and 91%, respectively, on CALFW and CPLFW. In regards to computational efficiency, AMD-FV achieves significant reductions in terms of the number of layers (32%), system parameters (26.92%) and FLOPs (41.91%), compared to ArcFace, positioning it as an attractive solution to real-world applications. A number of potential directions will be pursued as part of further research, including cross-domain applications, e.g., expand the use of AMD-FV in person re-identification, iris recognition, and fingerprint recognition, thus evaluating its adaptability. Moreover, we intend to combine the proposed system with other biometric modalities, for instance, gait, and consider its authentication accuracy in complex scenarios. Further research will evaluate its resilience against adversarial attacks to ensure security in such critical applications.
Supporting information
S1 Appendix.
Magnitude of Cross Product.
https://doi.org/10.1371/journal.pone.0324485.s001
References
1. 1. Ahonen T, Hadid A, Pietikäinen M. Face description with local binary patterns: application to face recognition. IEEE Trans Pattern Anal Mach Intell. 2006;28(12):2037–41. pmid:17108377
* View Article
* PubMed/NCBI
* Google Scholar
2. 2. Liu C, Wechsler H. Gabor feature based classification using the enhanced fisher linear discriminant model for face recognition. IEEE Trans Image Process. 2002;11(4):467–76. pmid:18244647
* View Article
* PubMed/NCBI
* Google Scholar
3. 3. Liu C, Wechsler H. Independent component analysis of Gabor features for face recognition. IEEE Trans Neural Netw. 2003;14(4):919–28. pmid:18238070
* View Article
* PubMed/NCBI
* Google Scholar
4. 4. Liu C. Gabor-based kernel PCA with fractional power polynomial models for face recognition. IEEE Trans Pattern Anal Mach Intell. 2004;26(5):572–81. pmid:15460279
* View Article
* PubMed/NCBI
* Google Scholar
5. 5. Cao Q, Shen L, Xie W, Parkhi OM, Zisserman A. VGGFace2: a dataset for recognising faces across pose and age. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018, pp. 67–74.
6. 6. Ahmed S, Srinivasu PN, Alhumam A, Alarfaj M. AAL and internet of medical things for monitoring type-2 diabetic patients. Diagnostics (Basel). 2022;12(11):2739. pmid:36359582
* View Article
* PubMed/NCBI
* Google Scholar
7. 7. Guleria P, Naga Srinivasu P, Ahmed S, Almusallam N, Alarfaj FK. XAI framework for cardiovascular disease prediction using classification techniques. Electronics. 2022;11(24):4086.
* View Article
* Google Scholar
8. 8. Liu W, Wen Y, Yu Z, Li M, Raj B, Song L. Sphereface: deep hypersphere embedding for face recognition. In: 2017 IEEE International Conference on Computer Vision (ICCV). 2017, pp. 6738–46.
9. 9. Wang H, Wang Y, Zhou Z, Ji X, Gong D, Zhou J, et al. CosFace: Large Margin Cosine Loss for Deep Face Recognition. 2018. pp. 5265–5274. Available from: https://ieeexplore.ieee.org/document/8578650/.
* View Article
* Google Scholar
10. 10. Deng J, Guo J, Xue N, Zafeiriou S. Arcface: additive angular margin loss for deep face recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2019, pp. 4690–9.
11. 11. Liu H, Zhu X, Lei Z, Li S. AdaptiveFace: adaptive margin and sampling for face recognition. In: 2019 IEEE International Conference on Image Processing (ICIP). 2019, pp. 11939–48.
12. 12. Zhang X, Zhao R, Qiao Y, Wang X, Li H. Adacos: adaptively scaling cosine logits for effectively learning deep face representations. IEEE Trans Pattern Anal Mach Intell. 2019:10815–24.
* View Article
* Google Scholar
13. 13. Li J, Xiao D, Lu T, Wei Y, Li J, Yang L. HAMFace: Hardness adaptive margin loss for face recognition with various intra-class variations. Expert Syst Appl. 2024;240:122384.
* View Article
* Google Scholar
14. 14. Gim T, Sohn KA. Regularization using noise samples identified by the feature norm for face recognition. IEEE Access. 2024.
* View Article
* Google Scholar
15. 15. Song Y, Wang F. CoReFace: sample-guided contrastive regularization for deep face recognition. Pattern Recognit. 2024;152:110483.
* View Article
* Google Scholar
16. 16. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016, pp. 770–8.
17. 17. Szegedy C, Ioffe S, Vanhoucke V, Alemi A. Inception-v4, inception-resnet and the impact of residual connections on learning. In: Proceedings of the AAAI conference on artificial intelligence, vol. 31. 2017. https://doi.org/10.1609/aaai.v31i1.11231
18. 18. Xie S, Girshick R, Dollár P, Tu Z, He K. Aggregated residual transformations for deep neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. IEEE; 2017, pp. 1492–500.
19. 19. Huang G, Liu Z, Van Der Maaten L, Weinberger K. Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. IEEE; 2017, pp. 4700–8.
20. 20. Chen Y, Li J, Xiao H, Jin X, Yan S, Feng J. Dual path networks. arXiv, preprint, 2017, pp. 4467–75.
21. 21. García J, Crosa P, Álvaro I, Pérez F, Alcocer P. Feature-preserving downsampling for medical images. 2015. Available from: https://upcommons.upc.edu/handle/2117/85388
22. 22. Sedlar K, Skutkova H, Vitek M, Provaznik I. Set of rules for genomic signal downsampling. Comput Biol Med. 2016;69:308–14. pmid:26078051
* View Article
* PubMed/NCBI
* Google Scholar
23. 23. Zhao G, Wang J, Zhang Z. Random shifting for CNN: a solution to reduce information loss in down-sampling layers. In: Proceedings of the Conference. 2017, pp. 3476–82.
24. 24. Kuen J, Kong X, Lin Z, Wang G, Yin J, See S. Stochastic downsampling for cost-adjustable inference and improved regularization in convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018, pp. 7929–38.
25. 25. Zhang Z, Gong X, Chen J. Face recognition based on adaptive margin and diversity regularization constraints. IET Image Processing. 2020;15(5):1105–14.
* View Article
* Google Scholar
26. 26. Yang Z, Zhu X, Jiang C, Liu W, Shen L. RamFace: race adaptive margin based face recognition for racial bias mitigation. Institute of Electrical and Electronics Engineers Inc.; 2021.
27. 27. Wang J, Zheng C, Yang X, Yang L. EnhanceFace: adaptive weighted softmax loss for deep face recognition. IEEE Signal Process Lett. 2022;29:65–9.
* View Article
* Google Scholar
28. 28. Khalifa A, Al-Hamadi A. JAMsFace: joint adaptive margins loss for deep face recognition. Neural Comput Applic. 2023;35(26):19025–37.
* View Article
* Google Scholar
29. 29. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016, pp. 770–8.
30. 30. Huang G, Liu Z, van der Maaten L, Weinberger K. Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017. 2261–9.
31. 31. Marsden J, Weinstein A. Exponentials and logarithms. Calculus I. 1985. pp. 307–35.
32. 32. de Freitas Pereira T, Anjos A, Marcel S. Heterogeneous face recognition using domain specific units. IEEE Trans Inform Forensic Secur. 2019;14(7):1803–16.
* View Article
* Google Scholar
33. 33. Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift. In: Proceedings of the 32nd International Conference on Machine Learning. 2015, pp. 448–56.
34. 34. Nair V, Hinton GE. Rectified linear units improve restricted Boltzmann machines. In: Proceedings of the 27th International Conference on Machine Learning. 2010, pp. 807–14.
35. 35. Xie S, Girshick R, Dollar P, Tu Z, He K. Aggregated residual transformations for deep neural networks. 2017, pp. 5987–95.
36. 36. Huang G, Ramesh M, Berg T, Learned-Miller E. Labeled faces in the wild: a database for studying face recognition in unconstrained environments. 2008.
37. 37. Kemelmacher-Shlizerman I, Seitz S, Miller D, Brossard E. The MegaFace benchmark: 1 million faces for recognition at scale. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016, pp. 4873–82.
38. 38. Whitelam C, Taborsky E, Blanton A, Maze B, Adams J, Miller T. Iarpa janus benchmark-b face dataset. 2017. Available from: http://ieeexplore.ieee.org/document/8014821/
* View Article
* Google Scholar
39. 39. Zheng T, Deng W, Hu J. Cross-age LFW: a database for studying cross-age face recognition in unconstrained environments. arXiv, preprint, 2017. https://arxiv.org/abs/1708.08197v1
40. 40. Zheng T, Deng W, Hu J. Cross-pose LFW: a database for studying cross-pose face recognition in unconstrained environments. 2018.
41. 41. Zhang K, Zhang Z, Li Z, Qiao Y. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process Lett. 2016;23(10):1499–503.
* View Article
* Google Scholar
42. 42. Schroff F, Kalenichenko D, Philbin J. Facenet: A unified embedding for face recognition and clustering. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2015, pp. 815–23.
43. 43. Zhang X, Fang Z, Wen Y, Li Z, Qiao Y. Range loss for deep face recognition with long-tailed training data. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017, pp. 5419–28.
44. 44. Sun Y, Liang D, Wang X, Tang X. DeepID3: face recognition with very deep neural networks. arXiv, preprint, 2015.
45. 45. Taigman Y, Yang M, Ranzato M, Wolf L. Deepface: closing the gap to human-level performance in face verification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2014, pp. 1701–8.
46. 46. Parkhi O, Vedaldi A, Zisserman A. Deep face recognition. In: BMVA British Machine Vision Conference. 2015, pp. 41.1-41.12.
47. 47. Wen Y, Zhang K, Li Z, Qiao Y. A discriminative feature learning approach for deep face recognition. In: Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, the Netherlands, October 11–14, 2016, Proceedings, Part VII. Springer. 2016, pp. 499–515.
48. 48. Xie W, Zisserman A. Multicolumn networks for face recognition. arXiv, preprint, 2018.
Citation: Khan ZA, Ahmed W, Liatsis P (2025) AMD-FV: Adaptive margin loss and dual path network+ for deep face verification. PLoS One 20(5): e0324485. https://doi.org/10.1371/journal.pone.0324485
About the Authors:
Zeeshan Ahmed Khan
Roles: Conceptualization, Data curation, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft
Affiliation: Department of Computer Science, School of Informatics, Xiamen University, Fujian, China
Waqar Ahmed
Roles: Methodology, Software, Visualization, Writing – review & editing
Affiliation: Department of Computer Science, Khalifa University, Abu Dhabi, United Arab Emirates
ORICD: https://orcid.org/0000-0001-5416-149X
Panos Liatsis
Roles: Conceptualization, Methodology, Project administration, Supervision, Validation, Writing – original draft, Writing – review & editing
E-mail: [email protected]
Affiliation: Department of Computer Science, Khalifa University, Abu Dhabi, United Arab Emirates
ORICD: https://orcid.org/0000-0002-5490-6030
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
1. Ahonen T, Hadid A, Pietikäinen M. Face description with local binary patterns: application to face recognition. IEEE Trans Pattern Anal Mach Intell. 2006;28(12):2037–41. pmid:17108377
2. Liu C, Wechsler H. Gabor feature based classification using the enhanced fisher linear discriminant model for face recognition. IEEE Trans Image Process. 2002;11(4):467–76. pmid:18244647
3. Liu C, Wechsler H. Independent component analysis of Gabor features for face recognition. IEEE Trans Neural Netw. 2003;14(4):919–28. pmid:18238070
4. Liu C. Gabor-based kernel PCA with fractional power polynomial models for face recognition. IEEE Trans Pattern Anal Mach Intell. 2004;26(5):572–81. pmid:15460279
5. Cao Q, Shen L, Xie W, Parkhi OM, Zisserman A. VGGFace2: a dataset for recognising faces across pose and age. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018, pp. 67–74.
6. Ahmed S, Srinivasu PN, Alhumam A, Alarfaj M. AAL and internet of medical things for monitoring type-2 diabetic patients. Diagnostics (Basel). 2022;12(11):2739. pmid:36359582
7. Guleria P, Naga Srinivasu P, Ahmed S, Almusallam N, Alarfaj FK. XAI framework for cardiovascular disease prediction using classification techniques. Electronics. 2022;11(24):4086.
8. Liu W, Wen Y, Yu Z, Li M, Raj B, Song L. Sphereface: deep hypersphere embedding for face recognition. In: 2017 IEEE International Conference on Computer Vision (ICCV). 2017, pp. 6738–46.
9. Wang H, Wang Y, Zhou Z, Ji X, Gong D, Zhou J, et al. CosFace: Large Margin Cosine Loss for Deep Face Recognition. 2018. pp. 5265–5274. Available from: https://ieeexplore.ieee.org/document/8578650/.
10. Deng J, Guo J, Xue N, Zafeiriou S. Arcface: additive angular margin loss for deep face recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2019, pp. 4690–9.
11. Liu H, Zhu X, Lei Z, Li S. AdaptiveFace: adaptive margin and sampling for face recognition. In: 2019 IEEE International Conference on Image Processing (ICIP). 2019, pp. 11939–48.
12. Zhang X, Zhao R, Qiao Y, Wang X, Li H. Adacos: adaptively scaling cosine logits for effectively learning deep face representations. IEEE Trans Pattern Anal Mach Intell. 2019:10815–24.
13. Li J, Xiao D, Lu T, Wei Y, Li J, Yang L. HAMFace: Hardness adaptive margin loss for face recognition with various intra-class variations. Expert Syst Appl. 2024;240:122384.
14. Gim T, Sohn KA. Regularization using noise samples identified by the feature norm for face recognition. IEEE Access. 2024.
15. Song Y, Wang F. CoReFace: sample-guided contrastive regularization for deep face recognition. Pattern Recognit. 2024;152:110483.
16. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016, pp. 770–8.
17. Szegedy C, Ioffe S, Vanhoucke V, Alemi A. Inception-v4, inception-resnet and the impact of residual connections on learning. In: Proceedings of the AAAI conference on artificial intelligence, vol. 31. 2017. https://doi.org/10.1609/aaai.v31i1.11231
18. Xie S, Girshick R, Dollár P, Tu Z, He K. Aggregated residual transformations for deep neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. IEEE; 2017, pp. 1492–500.
19. Huang G, Liu Z, Van Der Maaten L, Weinberger K. Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. IEEE; 2017, pp. 4700–8.
20. Chen Y, Li J, Xiao H, Jin X, Yan S, Feng J. Dual path networks. arXiv, preprint, 2017, pp. 4467–75.
21. García J, Crosa P, Álvaro I, Pérez F, Alcocer P. Feature-preserving downsampling for medical images. 2015. Available from: https://upcommons.upc.edu/handle/2117/85388
22. Sedlar K, Skutkova H, Vitek M, Provaznik I. Set of rules for genomic signal downsampling. Comput Biol Med. 2016;69:308–14. pmid:26078051
23. Zhao G, Wang J, Zhang Z. Random shifting for CNN: a solution to reduce information loss in down-sampling layers. In: Proceedings of the Conference. 2017, pp. 3476–82.
24. Kuen J, Kong X, Lin Z, Wang G, Yin J, See S. Stochastic downsampling for cost-adjustable inference and improved regularization in convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018, pp. 7929–38.
25. Zhang Z, Gong X, Chen J. Face recognition based on adaptive margin and diversity regularization constraints. IET Image Processing. 2020;15(5):1105–14.
26. Yang Z, Zhu X, Jiang C, Liu W, Shen L. RamFace: race adaptive margin based face recognition for racial bias mitigation. Institute of Electrical and Electronics Engineers Inc.; 2021.
27. Wang J, Zheng C, Yang X, Yang L. EnhanceFace: adaptive weighted softmax loss for deep face recognition. IEEE Signal Process Lett. 2022;29:65–9.
28. Khalifa A, Al-Hamadi A. JAMsFace: joint adaptive margins loss for deep face recognition. Neural Comput Applic. 2023;35(26):19025–37.
29. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016, pp. 770–8.
30. Huang G, Liu Z, van der Maaten L, Weinberger K. Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017. 2261–9.
31. Marsden J, Weinstein A. Exponentials and logarithms. Calculus I. 1985. pp. 307–35.
32. de Freitas Pereira T, Anjos A, Marcel S. Heterogeneous face recognition using domain specific units. IEEE Trans Inform Forensic Secur. 2019;14(7):1803–16.
33. Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift. In: Proceedings of the 32nd International Conference on Machine Learning. 2015, pp. 448–56.
34. Nair V, Hinton GE. Rectified linear units improve restricted Boltzmann machines. In: Proceedings of the 27th International Conference on Machine Learning. 2010, pp. 807–14.
35. Xie S, Girshick R, Dollar P, Tu Z, He K. Aggregated residual transformations for deep neural networks. 2017, pp. 5987–95.
36. Huang G, Ramesh M, Berg T, Learned-Miller E. Labeled faces in the wild: a database for studying face recognition in unconstrained environments. 2008.
37. Kemelmacher-Shlizerman I, Seitz S, Miller D, Brossard E. The MegaFace benchmark: 1 million faces for recognition at scale. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016, pp. 4873–82.
38. Whitelam C, Taborsky E, Blanton A, Maze B, Adams J, Miller T. Iarpa janus benchmark-b face dataset. 2017. Available from: http://ieeexplore.ieee.org/document/8014821/
39. Zheng T, Deng W, Hu J. Cross-age LFW: a database for studying cross-age face recognition in unconstrained environments. arXiv, preprint, 2017. https://arxiv.org/abs/1708.08197v1
40. Zheng T, Deng W, Hu J. Cross-pose LFW: a database for studying cross-pose face recognition in unconstrained environments. 2018.
41. Zhang K, Zhang Z, Li Z, Qiao Y. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process Lett. 2016;23(10):1499–503.
42. Schroff F, Kalenichenko D, Philbin J. Facenet: A unified embedding for face recognition and clustering. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2015, pp. 815–23.
43. Zhang X, Fang Z, Wen Y, Li Z, Qiao Y. Range loss for deep face recognition with long-tailed training data. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017, pp. 5419–28.
44. Sun Y, Liang D, Wang X, Tang X. DeepID3: face recognition with very deep neural networks. arXiv, preprint, 2015.
45. Taigman Y, Yang M, Ranzato M, Wolf L. Deepface: closing the gap to human-level performance in face verification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2014, pp. 1701–8.
46. Parkhi O, Vedaldi A, Zisserman A. Deep face recognition. In: BMVA British Machine Vision Conference. 2015, pp. 41.1-41.12.
47. Wen Y, Zhang K, Li Z, Qiao Y. A discriminative feature learning approach for deep face recognition. In: Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, the Netherlands, October 11–14, 2016, Proceedings, Part VII. Springer. 2016, pp. 499–515.
48. Xie W, Zisserman A. Multicolumn networks for face recognition. arXiv, preprint, 2018.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2025 Khan et al. This is an open access article distributed under the terms of the Creative Commons Attribution License: http://creativecommons.org/licenses/by/4.0/ (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Face verification is important in a variety of applications, for instance, access control, surveillance, and identification. Existing methods often struggle with the challenges of dataset imbalance and manual hyperparameter tuning. To address this, we propose the Adaptive Margin Loss and Dual Path Network+ (AMD-FV) for deep face verification. Two innovations are introduced, namely, Adaptive Margin Loss (AML) and Dual Path Network+ (DPN+). AML aims at automating the selection of margin and scale hyperparameters in large margin loss functions, thus, eliminating the need for manual tuning. Input dissimilarity information is used to estimate the margin, while the scale parameter is computed using the number of classes and AML’s range. Next, DPN+ enhances the original Dual Path Network by redesigning the first block with a series of 3x3 convolutions, batch normalization, and ReLU activations, leveraging shared connections across layers, leading to increases in spatial resolution and computational cost efficiency, while maximizing the use of discriminative features. We present comprehensive experiments on five diverse face verification datasets (LFW, Megaface, IJB-B, CALFW, and CPLFW) to demonstrate the effectiveness of the proposed approach. The results show that AMD-FV outperforms state-of-the-art methods, achieving a verification accuracy of 99.75% on LFW, improving the True Acceptance Rate by 6% on IJB-B at a False Acceptance Rate of 0.001, compared to VGGFace2, and attaining a Rank-1 identification score of 92.16% on Megaface, surpassing the CosFace model by 9.44%.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer