Multi Task Impersonation Prevention Face Recognition Network (MTIPFRNet) and Layer Wise Triplet Loss (Record no. 16591)

MARC details
000 -LEADER
fixed length control field 04514nam a22002057a 4500
082 ## - DEWEY DECIMAL CLASSIFICATION NUMBER
Classification number 006.4
Item number U721M
100 ## - MAIN ENTRY--AUTHOR NAME
Personal name Surana, Sushil Kumar
245 ## - TITLE STATEMENT
Title Multi Task Impersonation Prevention Face Recognition Network (MTIPFRNet) and Layer Wise Triplet Loss
Statement of responsibility, etc by Sushil Kumar Surana
260 ## - PUBLICATION, DISTRIBUTION, ETC. (IMPRINT)
Place of publication IIT Jodhpur
Name of publisher Department of Computer Science and Technology
Year of publication 2023
300 ## - PHYSICAL DESCRIPTION
Number of Pages x,45p.
Other physical details HB
500 ## - GENERAL NOTE
General note Recent work in the field of face recognition aims at achieving high accuracy, and they have been succeeding. Some of the face recognition algorithms have been able to achieve accuracy that surpasses even human beings for unseen data. Some recent literary works have focused on face identification and recognition even in cases of obstructions on the face, which could be intentional or unintentional, such as masks, makeup, etc.<br/><br/>In this work, we propose a novel multi-task framework for face recognition with the aim of increasing the capability of AI systems to match faces and reduce tolerance toward impersonation due to its critical applicability in areas requiring high secrecy, authenticity, and confidentiality. We propose to divide the face into several regions such as the mouth region, periocular region, nose region, etc., along with using the capability of existing systems that work on features identified from the complete face. The proposed framework takes four primary sub-tasks, namely Facial Feature matching, Periocular feature matching, Nose feature matching, and Mouth feature matching. Differently trained models of similar architecture for each of these features have been used. For the purpose of final authentication or verification, distance is calculated between each of the feature embeddings. Then these four distances are fed into an ensemble of classifiers to make the final decision on the match of the face. The proposed pipeline is able to attain an accuracy of 90.4%, which is approximately 1% higher than the best state-of-the-art accuracy for the under-consideration private dataset. We have also carried out an ablation study on this dataset, and the findings suggest a minimum of 0.8% points drop in case of the removal of any of the four feature embeddings. Moreover, the removal of any of these four tasks leads to an increase in the accuracy gap of matching and non-matching pairs, i.e., such removal induces bias.<br/><br/>This research work also proposes Layer-Wise Triplet Loss considering the class-independent nature of Triplet loss and contrastive losses. The newly proposed approach trains a neural network with loss for each layer. The loss for any layer impacts that layer and all layers preceding it. In this way, every layer is learning not only from the final output of the network but also from all internal layers as well. This learning with all internal layers and respective losses expedites the learning, and the model is trained quickly in very few epochs. In any deep neural network, lower layers contain better localization properties, and deeper layers turn to be more class-specific. In this layer-wise training, the same continues; however, since the proposed approach uses the loss functions which are independent of class representation, each layer is expected to generate features that are more representative for the final decision in verification or matching tasks. In order to avoid overfitting with the layer-wise loss function, we recommend a dropout of around 50% for each layer.<br/><br/>Random variations in brightness, color jitters, rotation, and noise induction will also help in avoiding overfitting and increasing the generalizability of the model. Increasing weights with the depth of layer-wise loss ensures that the lower layer of the model not only learns from its outputs but also learns from the outputs of higher layers. Experiments indicate that the newly proposed training approach, using the layer-wise loss function, is able to achieve accuracy comparable with the same state-of-the-art technology and is more generalizable in very few training iterations. The proposed approach saves not only training time but also training resources and hence is more environment-friendly.
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical Term Department of Computer Science and Technology
Topical Term Face Recognition Algorithms
Topical Term Facial Feature Matching
Topical Term Deep Neural Networks in Face Recognition
Topical Term MTech Theses
700 ## - ADDED ENTRY--PERSONAL NAME
Personal name Vatsa, Mayank
Personal name Singh, Richa
942 ## - ADDED ENTRY ELEMENTS (KOHA)
Koha item type Thesis
Holdings
Withdrawn status Lost status Damaged status Not for loan Collection code Permanent Location Current Location Shelving location Date acquired Source of acquisition Full call number Accession Number Price effective from Koha item type
        Theses S. R. Ranganathan Learning Hub S. R. Ranganathan Learning Hub Reference 2024-04-01 Office of Academics 006.4 U721M TM00527 2024-07-01 Thesis