iconOpen Access

ARTICLE

crossmark

Iris Liveness Detection Using Fragmental Energy of Haar Transformed Iris Images Using Ensemble of Machine Learning Classifiers

Smita Khade1, Shilpa Gite1,2,*, Sudeep D. Thepade3, Biswajeet Pradhan4,5,*, Abdullah Alamri6

1 Symbiosis International (Deemed University), Symbiosis Institute of Technology, Pune, 412115, India
2 Symbiosis Centre for Applied Artificial Intelligence, Symbiosis International (Deemed University), Pune, 412115, India
3 Computer Engineering, Pimpri Chinchwad College of Engineering, Pune, 411044, India
4 Centre for Advanced Modelling and Geospatial Information Systems (CAMGIS), School for Civil and Environmental Engineering, Faculty of Engineering and Information Technology, University of Technology Sydney, Sydney, New South Wales, 2007, Australia
5 Earth Observation Center, Institute of Climate Change, University Kebangsaan Malaysia, UKM, Bangi, Selangor, 43600, Malaysia
6 Department of Geology & Geophysics, College of Science, King Saud University, P.O. Box 2455, Riyadh, 11451, Saudi Arabia

* Corresponding Authors: Shilpa Gite. Email: email; Biswajeet Pradhan. Email: email

(This article belongs to this Special Issue: AI-Driven Engineering Applications)

Computer Modeling in Engineering & Sciences 2023, 136(1), 323-345. https://doi.org/10.32604/cmes.2023.023674

Abstract

Contactless verification is possible with iris biometric identification, which helps prevent infections like COVID-19 from spreading. Biometric systems have grown unsteady and dangerous as a result of spoofing assaults employing contact lenses, replayed the video, and print attacks. The work demonstrates an iris liveness detection approach by utilizing fragmental coefficients of Haar transformed Iris images as signatures to prevent spoofing attacks for the very first time in the identification of iris liveness. Seven assorted feature creation ways are studied in the presented solutions, and these created features are explored for the training of eight distinct machine learning classifiers and ensembles. The predicted iris liveness identification variants are evaluated using recall, F-measure, precision, accuracy, APCER, BPCER, and ACER. Three standard datasets were used in the investigation. The main contribution of our study is achieving a good accuracy of 99.18% with a smaller feature vector. The fragmental coefficients of Haar transformed iris image of size 8 * 8 utilizing random forest algorithm showed superior iris liveness detection with reduced featured vector size (64 features). Random forest gave 99.18% accuracy. Additionally, conduct an extensive experiment on cross datasets for detailed analysis. The results of our experiments show that the iris biometric template is decreased in size to make the proposed framework suitable for algorithmic verification in real-time environments and settings.

Graphical Abstract

Iris Liveness Detection Using Fragmental Energy of Haar Transformed Iris Images Using Ensemble of Machine Learning Classifiers

Keywords


1  Introduction

Automatic human access to a system has become relatively simple in the digital era. Confirmation of the user’s identification is critical for automated system access. Biometric authentication systems employ biometric features to confirm the identification of a user [1]. Compared to conventional password-based traditional verification systems, the biometric system has a distinct advantage. It minimizes the need to memorize a passcode, pin, or keep a card in hand [2]. Biometric authentication can be thought of as an additional layer of authentication for security-critical cyber applications and existing traditional authentication procedures. Today, there are indeed different businesses for biometric systems. The majority of the sectors seem to be rising quickly. According to www.statista.com (accessed on 24th June 2022), the business for contactless biometric technologies is forecasted to increase by roughly 30.15 billion US dollars by 2027, while the overall biometric sciences industry is forecasted to touch 19.08 billion US dollars in 2021 [3]. Iris is widely employed in the verification and validation of people. In most applications, it uses, because of its complex textures [4] and distinctive features, such as the UIDAI project for citizen identification in India, the Amsterdam airport, and the Canada-US border onon-US [5].

In comparison to fingerprint and face authentication, iris authentication delivers an additional steadfast contactless user verification. The contactless method aids in the prevention of diseases like COVID-19 [6]. Despite the iris having a distinct textural structure, the imposter might falsify it [7]. Table 1 shows the iris presentation attacks used that are found in the literature [8].

images

Individuals frequently assault the biometric system to get admittance to another person’s credentials or to conceal their accurate individuality. The iris identification system can be readily fooled by means of alternative contact lenses (which can be transparent, textured, colored) [10], replaying the video, or using a print attack [9]. As a result, understanding the risk and susceptibility is critical for safeguarding the biometric system [14]. The complicated risk of biometric spoofing is minimized by assessing the liveness of biometric features prior to authentication [15]. The main objective of this study is to identify iris liveness detection with reduced feature vectors.

The main objective of this study is to identify iris liveness detection with reduced feature vectors. Following are the novelty and the main contributions of this paper.

•   Initiatory utilization of ‘fragmental coefficients of Haar transformed iris image data’ as signatures in iris liveness detection;

•   Determining the smallest size of fragmental coefficients that might be used for feature generation without impairing iris liveness detection performance;

•   To identify which classifier is optimal in iris liveness detection, the performance of machine learning (ML) classifiers and their ensemble combinations are analyzed;

•   Testing the feasibility of the developed iris liveness detection method against a variety of existing benchmark datasets.

The paper’s organization is presented herewith. Section 2 elaborates on an outline of existing methodologies. Section 3 portrays the proposed approach to iris liveness detection. The experimentation setup is put forth in Section 4. While Section 5 elaborates on the noted outcomes and the conclusions taken from the findings. Section 6 is where the discussion takes place. Finally, final thoughts, limitations, and future research suggestions are presented in Section 7.

2  Prevailing Iris Liveness Detection Techniques

Various strategies have been adopted to determine whether the acquired biometric traits are alive prior to the authentication. Several of the most well-known techniques are addressed in this section.

For liveness identification, Agarwal et al. [16] explored fingerprints with iris. To create a fingerprint vector function, the essential Haralick statistical characteristics use GLCM and NGTDM. The iris texture feature is utilized to improve the device’s performance. To evaluate if this model is more efficient than the current one, Agarwal used a standard dataset. GLCM has an extensive feature vector size in the current system. Iris spoofing attacks are detected using rotation-invariant features of Polar harmonic transformations and Zernike moments [5]. Spoofing assaults on numerous sensors significantly impact the system’s overall competence. The system detects attacks like iris print and contact lenses.

Thavalengal et al. created a that uses smartphones to take RGB along with NIR images of the iris and eye [17]. Identification is made using pupil localization techniques and distance measures. 4096-dimensional features are examined for feature vector generation, which is a considerable number. Although the author claims a high rate of liveness recognition, he does not work with standard datasets. Authors Fathy et al. have not examined the segmentation or normalization processes commonly utilized in Iris liveness discovery systems [13]. The original image is broken down into wavelets using Wavelet Packets (WPs). Although the author claims it is 100 percent accurate, it is not working with all genres of assaults and only covers a few spoof attacks.

Iris liveness detection will be made utilizing regional features by author Hu et al. [18]. The interaction of the properties of nearby regions is used to create regional traits. The author utilized one hundred forty-four relational measurements based on regional attributes during the experiment.

Using pupil dynamics, the author Czajka [19] created the liveness identification system. The pupil reaction is tested in this system using rapid changes in the intensity of light. In the case of the eye reacting to the variations in light intensity, it is alive; otherwise, presentation attack image. In [19], non-linear and linear SVM is employed to categorize natural reactions and impulsive oscillations. The system’s shortcoming is that it measures a variety of functions that take time. There are inaccuracies in the observation because the data utilized in this research does not include any details from elderly adults.

Author Fang et al. [20] apply many data augmentation methods to generate variability. The strategy-level and the score-level combination of fusion methods are used for Iris PAD. Bassi et al. [21] detected PAD using NIR, visible domain, cross-datasets and cross-spectrum datasets. Authors concluded that Cross-PA and cross-datasets are still challenging, as EER values above 20% in most of cases.

A technique to detect Accurate Ocular Regions was created by Naqvi et al. [22]. This solution uses deep neural network variants. The system’s evaluation considers publicly available databases. Kimura et al. [23] developed a CNN-based liveness detection System that improves model accuracy by modifying hyperparameters. “Attack Presentation Classification Error Rate [APCER]” and “Bonafede Presentation Classification Error Rate [BPCER]” metrics are taken to assess the system’s performance. The hyperparameters are all studied in this work. This method is solely effective against prints and contact lens attacks. Author [24] studied multiple transfer learning models to detect the iris liveness and concluded that EfficeintnetB7 gives highest classification accuracy.

Only a few studies were found to be robust against all sorts of spoofing assaults [3,25]. Most of the studies used a higher size feature vector. Based on these findings, it can be believed that there is a necessity for a classifier or ensembles for the detection of every sort of spoofing assault.

3  Proposed Iris Liveness Detection Utilizing Fragmental Energy of Haar Transformed Iris Images

The iris recognition system is prone to a range of security threats. Because of these flaws, the system is less trustworthy for robust authentication applications. The study employs fragmental energy of Haar modified iris images to attempt iris liveness detection. These fragmental energies were employed as features to detect whether the iris was real or fake. Because of these characteristics, the suggested methodology does not require any pre-processing, such as segmentation, normalization, or localization, which are commonly employed by methods presented in the literature. These fragmental energies were employed as features to detect whether the iris was real or fake. Because of these characteristics, the suggested methodology does not require any pre-processing, such as segmentation, normalization, or localization, which are commonly employed by methods presented in the literature. These fragmental energies make the suggested technique faster and more accessible [26]. Resizing the iris image to 256 * 256 is the sole pre-processing performed in the proposed framework. The iris liveness detection process depicted in Fig. 1 is a block diagram. There are three phases in the proposed system. Resizing [pre-processing] of iris images, feature formation, and classification with iris liveness identification.

images

Figure 1: Block diagram of the projected iris liveness detection employing fragmental energy of Haar transformed iris images

3.1 Pre-Processing

The importance of iris pre-processing in iris liveness detection cannot be overstated. Two iris pre-processing techniques are used in the suggested algorithm. Because images are obtained using three standard datasets, each dataset stores images of different sizes. We normalized the original 256 * 256 images in pre-processing to ensure they remained intact throughout the experiment. At the same time, photographing various datasets with various sensors, some (LG, Content, Vista) acquired images in RGB format, while others [LG, Dalsa] acquired grayscale images. The images were then converted to grayscale to keep their originality.

3.2 Feature Formation with the Fragmental Energy of Transformed Iris

A scaled iris image is subjected to the Haar transform. The Haar transform allows content with high energy to congregate in the transform domain’s lower frequency section [27]. The Haar coefficients are described as below:

In case s = 0, the function of Haar is presented as Eq. (1).

ho(t)=1/N(1)

In case s > 0, the function of Haar is presented by Eq. (2).

hs(t)=1/N{2p22p20}(2)

The nonzero part of the function’s amplitude and width are determined by p, whereas the nonzero part of the function’s placement is determined by q.

In the Haar transform iris image, the left highest corner has the higher energy and crucial information, as shown in Fig. 2. This results in considerable energy compression in a limited count of high energy coefficients. As a result, these are the preferred feature vector elements. To construct feature vectors for proposed iris liveness detection, 256 * 256, 128 * 128, 64 * 64, 32 * 32, 16 * 16, 8 * 8, and 4 * 4. Pixels are used to capture the high-energy portion of Haar transformed iris image coefficients.

images

Figure 2: Proposed fragmental energy-based feature creation approach for liveness identification from cosine transformed iris images

These feature vectors support the reduction of the size of feature vectors. As a result, iris-liveness detection is speedier. The compacted high energy improves iris liveness detection accuracy in these low-frequency coefficients. These high-energy features are then employed for training the ML models working to detect iris liveness.

3.3 Iris Liveness Detection Using Meachine Learning Classifiers

The suggested method employs a combination of machine learning (ML) classifiers and ensembles. Naive Bayes (NB), Support Vector Machine (SVM), Random Forest (RF), and J48 are the ML classifiers [15] used here, along with SVM+NB+RF SVM+RT+RF, SVM+MLP+RF ensembles of classifiers.

Ensemble method—Using multiple models concurrently on a single set for categorization is always preferable to just one model. Ensemble learning is the name for this technique [23]. Different classifiers are used to train a model, and the end output is an ensemble of the classifiers. The suggested method employs majority voting logic for an ensemble of ML classifiers.

These classifiers are trained using a tenfold cross-validation approach. The most effective method to train ML classifiers is tenfold cross-validation. Tenfold cross-validation allows all data in the dataset to be considered as either test or training data, giving a more unbiased classifier. The ensembles of ML classifiers are created using the majority voting mechanism.

Any deep learning architecture to perform well needs a considerable amount of data. Due to this, the time complexity increase. This disadvantage was overcome with the help of ML classifiers with handcrafted feature extraction. In the case of ensembles of classifiers, multiple classifiers help to classify correctly with majority vote logic. So, with the help of combined ML classifiers, the performance is superior compared to deep learning architectures/CNN.

4  Experimentation Setup

The investigative results of the proposed method are discussed in this section. The experiments were performed using an Intel (R) Core (TM) i3-6006U CPU @ 2.0 GHz, 12 GB RAM, and 6a 464-bit operating system with MATLAB R2015a as a programming platform. Clarkson LiveDet2013 (Clarkson 2013), LiveDet2015 (Clarkson 2015), and IITD Combined Spoofing datasets (IIITD CSD) were used to explore the suggested approach to iris liveness detection.

4.1 Description of Datasets

Three publicly available benchmark datasets are taken in this investigation. The dataset’s detailed description is as follows:

4.1.1 Clarkson LivDet2013

Around 1356 iris images are included in the Clarkson 2013 dataset [28]. There are two sets of data in this dataset: testing and training. The Dalsa sensor is utilized to acquire the iris. The images from training data (as given by the data creator) are used in this study for training purpose, and testing images (as given by the data creator) are used for testing purposes purpose. The dataset, sensors utilized in image acquisition, and the Count of images taken for training and testing a model during this exploration, with some examples, are all listed in Table 2.

images

4.1.2 Clarkson LivDet2015

LivDet2015—Images used in LivDet2015 dataset are captured using Dalsa and LG sensors [29]. Images are divided into three categories: live, printed, and pattern. In total, 25 subjects are used for live, images and patterns are printed; 15 subjects are used. The whole dataset is partitioned into training and testing. The images from training data (as given by the data creator) are used in this study for training purposes, and testing images (as given by the data creator) are used for testing purposes. The dataset, sensors utilized in image acquisition, and the count of images taken for training and testing a model during this exploration, with some examples, are all listed in Table 3.

images

4.1.3 IIITD Combined Spoofing Database [IIITD CSD]

Two iris detectors, a Cogent sensor, and a Vista iris sensor, were utilized to create the images used in this collection [30,31].

Normal, Print-Capture attack, and Print-Scan attack are the three types of images available in the dataset [32]. A 60:40 ratio is used for training testing split. The sensors utilized in image acquisition and the Count of images taken during this exploration with some sample images are all listed in Table 4.

images

4.2 Performance Measures

F-measure, accuracy, recall, precision, ACER (Average Classification Error Rate), APCER (Attack Presentation Classification Error Rate), and BPCER (Bonafide Presentation Classification Error Rate) are employed as performance metrics utilized here. Let the true positive, true negative, false positive, and false negative of the iris liveness detection be TP, TN, FP, and FN, respectively. The TP designates projected authentic data instances, which are truly what they are. The TN returns data examples that have been identified as spoofed and are also spoofed examples [2]. FP denotes that the examples were detected as bonafide but were spoofed. The data examples were detected as presentation attacks imaged, but bonafide iris examples are shown in FN. The formulas for the performance metric utilized are given by Eqs. (3) to (9).

Accuracy=(TP+TN)TP+TN+FP+FN(3)

Precision=TPTP+FP(4)

Recall=TPTP+TN(5)

FMeasures=2[PrecisionRecall][Precision+Recall](6)

APCER=FPTN+FP(7)

BPCER=FNTP+FN(8)

ACER=(APCER+NPCER)2(9)

5  Experimentation Setup

The benchmark datasets for all feature size variants are taken to test the proposed iris liveness detection method. Performance measurements such as accuracy, F-measure, precision, and recall are considered for testing versions of the proposed iris liveness detection technique.

5.1 Clarkson LivDet2013 Results

Fig. 3 compares the performance of the investigated fragmental coefficients for a particular ML classifier in the proposed technique of iris liveness detection, which was evaluated on the Clarkson 2013 dataset. In Fig. 3, it can be seen that fragmental coefficients 8 * 8 outperformed other fragmental coefficient combinations for all classifiers. The highest noted iris liveness detection accuracy comes around 98.10%, with 8 * 8 fragmental coefficients using a RF classifier.

images

Figure 3: Performance assessment of considered fragmental coefficients for specific ML classifiers in iris liveness detection for Clarkson 2013 dataset

From Fig. 4, it has been noted that the performance improves as the size of the feature vector is reduced from 256 * 256 to 8 * 8 and then begins to deteriorate with feature vector size 4 * 4. This demonstrates that the fragmental coefficients of Haar transformed iris images provide more outstanding iris liveness recognition capabilities while maintaining a small feature vector size, proving the importance of the suggested method. The highest average accuracy, 95.94% achieved by 16 * 16 fragmental coefficients.

images

Figure 4: The performance assessment for the Clarkson 2013 dataset by averaging the specific fragmental coefficients in iris liveness detection

Table 5 appraises the performance of specific ML classifiers and ensembles of classifiers for iris liveness detection tested on the Clarkson 2013 dataset. From Table 5, it can be noted that NB classifiers give the highest average ACER, whereas the lowest average ACER, 2.63% achieved by SVM+MLP+RF ensembles of classifiers.

images

5.2 IIITD Combined Spoofing Database Results

Fig. 5 reflects the performance comparison of considered Fragmental coefficients for particular ML classifiers in the proposed iris liveness detection, tested on IIITD CSD. Here, it is noted that, for all classifiers, 4 * 4 fragmental coefficients outperformed other fragmental coefficient combinations for IIITD CSD. The highest accuracy, 97.94% achieved by using an RF classifier. Because of its significant energy compaction, Haar can produce the best results with 4 * 4 fragmental coefficients.

images

Figure 5: Performance assessment of considered fragmental coefficients for specific ML classifiers in iris liveness detection for IIITD CSD dataset

From Fig. 6, it can be seen that the performance improves as the feature vector size is compacted from 256 * 256 to 8 * 8 and then begins to deteriorate with feature vector size 4 * 4. This demonstrates that the fragmental coefficients of Haar transformed iris images provide more excellent iris liveness recognition capabilities while maintaining a small feature vector size. The highest average accuracy, 96.80% achieved by 8 * 8 fragmental coefficients.

images

Figure 6: Performance assessment by averaging the specific fragmental coefficients in iris liveness detection for IIITD CSD

Table 6 provides the performance assessment of specific ML classifiers and ensembles of classifiers in the projected iris liveness detection explored on the IIITD CSD dataset. It is noted from the table that Decision Tree (J48) classifiers give the highest average ACER, whereas the lowest average ACER, 2.29%, is achieved by SVM+NB+RF ensembles of classifiers. The majority voting technique generates ensembles of classifiers, so they provide the best classification accuracy.

images

5.3 Clarkson 2015

Fig. 7 reflects the performance comparison of considered Fragmental coefficients for particular ML classifiers in proposed iris liveness detection, tested on the Clarkson 2015 dataset. Here, it is noted that for all classifiers, 16 * 16 fragmental coefficients outperformed other fragmental coefficient combinations for Clarkson 2015 dataset. The highest accuracy, 99.90% achieved by using an SVM+RF+RT ensemble classifier. Because of its significant energy compaction, Haar can produce the best results with 8 * 8 fragmental coefficients.

From Fig. 8, it has been noted that the performance improves as the feature vector size is compacted from 256 * 256 to 8 * 8 and then begins to deteriorate with feature vector size 4 * 4. This demonstrates that the fragmental coefficients of Haar transformed iris images provide more excellent iris liveness recognition capabilities while maintaining a small feature vector size. The highest average accuracy, 99.58% achieved by 8 * 8 fragmental coefficients.

images

Figure 7: Performance assessment of considered fragmental coefficients for specific ML classifiers in iris liveness detection for Clarkson 2015 dataset

images

Figure 8: Performance assessment by averaging the specific fragmental coefficients in iris liveness detection for Clarkson 2015 dataset

Table 7 gives the performance assessment of specific ML classifiers and ensembles of classifiers in the projected iris liveness detection explored on the Clarkson 2015 dataset. It is noted from the table that Decision Tree (J48) classifiers give the highest average ACER, whereas the lowest average ACER, 3.2%, is achieved by SVM+RF+MLP ensembles of classifiers. The majority voting technique generates ensembles of classifiers, so they provide the best classification accuracy.

images

5.4 Cross Datasets Evaluation Results

In this section, the results of cross datasets performances are explained in detail [33]. The first scenario, where model train on Clarkson 2015 datasets was evaluated on the Clarkson 2013 and IIITD test datasets. The evaluation results are presented in Table 8, where bold digits indicate the highest accuracy. From Table 8, we observed that Clarkson’s 2015 datasets give a lower ACER of nearly zero percent. However, Clarkson’s 2013 datasets do not perform well and give a high ACER of around 74%.

images

In this section, the results of cross datasets performances are explained. The first scenario, where model train on Clarkson 2015 datasets was evaluated on the Clarkson 2013 and IIITD test datasets. The evaluation results are presented in Table 8, where bold digits indicate the highest accuracy. From Table 8, we observed that Clarkson 2015 gives a lower ACER of nearly zero percentage. However, Clarkson 2013 datasets do not perform well and give a high ACER of around 74%.

Fig. 9 shows performance evaluation on cross datasets. It can be seen that our model outperforms in IIITD and Clarkson 2015 datasets, however, shows low performances for Clarkson 2013 dataset. One possible reason for this is Clarkson 2013 dataset has a smaller number of images compared to the other two datasets.

images

Figure 9: Performance assessment on cross datasets evaluation

Table 9 represents the performance comparison of fragmental coefficients across all datasets used for implementation with an average of percent accuracy, percent precision, percent recall, and percent F-ratio values. The highest performance is represented in bold and underlined. From Table 8, it can be seen that reducing the number of higher energy coefficients from 128 * 128 to 8 * 8 improves performance since the common part is reduced and discriminative is emphasized more.

images

6  Discussion

The proposed experiment was performed using the Haar transform. The fundamental goal of experimenting with the transform domain is to learn more about how the image is split into low and high-energy parts, reducing the feature vector size and speeding up retrieval. The Haar statistic is used to transform data [27].

By applying Haar transform on Iris images, high energy coefficients of transformed iris images sized 256 * 256, 128 * 128, 64 * 64, 32 * 32, 16 * 16, 8 * 8, and 4 * 4 do generate feature vectors for the projected iris-liveness detection. The procedures outlined in Section 3.2 are used to create the feature vector. Seven distinct ML and ensembles of classifiers are trained using these features. These classifiers are trained using the tenfold cross-validation method to detect presentation attacks. Three benchmark datasets are taken for testing: Clarkson 2013, Clarkson 2015, and the IIITD combined spoofing database. These three datasets explain in Section 4.1. Accuracy, Precision, Recall, and F-ratio and ISO standard metrics APCER, BPCER, and ACER are utilized to compare the performance of all the variants of the suggested approach. Section 4.2 describes several performance measures.

As stated in Sections 5.1 and 5.2, feature extraction using Haar has shown outstanding average classification accuracy. For the Clarkson 2013 dataset, the highest noted iris liveness detection accuracy comes around 98.10%, with 8 * 8 fragmental coefficients obtained using the classifier RF. The highest average accuracy, 95.94% was achieved by 16 * 16 fragmental coefficients, whereas the maximum average accuracy was 96.31%, and the average ACER was around 2.63%, achieved by SVM+MLP+RF ensembles of classifiers. For Clarkson 2015, the highest accuracy achieved was 99.90% by using an SVM+RF+RT ensemble classifier. For IIITD CSD, the uppermost accuracy of 97.94% was obtained by using the RF classifier. The uppermost average accuracy was 96.80% achieved by 8 * 8 fragmental coefficients, whereas the utmost average accuracy was 97.05%, and average ACER was around 2.29%, achieved by SVM+NB+RF ensembles of classifiers. The findings show that our suggested approach distinguishes between the bonafide and presentation attack images artifacts using the Haar transform approach. Table 10 shows a comparison of the suggested strategy to recent studies in this area.

images

Compared to similar current techniques based on fragmented energy, the Haar transformation better discerns between real and artificial artifacts. The results reveal that the proposed method reduces classification error and gets better accuracy when compared to earlier ways of detecting presentation attacks using an iris liveness detection. Table 10 summarizes this information. The proposed strategy outperforms some recent existing studies. As many recent studies used the different train and test datasets, so we partition Table 10 into two parts, comparing with same datasets and different datasets. Even though some studies outperform, our approach achieved this performance with reduced feature vector size (only with 64 features).

While implementing this study, we faced a few challenges-one of them was getting access to datasets with permissions for use in experimentation. The second challenge we faced was the explorations of the proposed method with the number of test runs was time-consuming task.

The limitations of the study are: (i) it was applied only two pre-processing techniques like resizing and converting an image into grayscale; (ii) this experiment is limited to image size 256 by 256 and only grayscale images were used during this study.

7  Conclusion

The paper proposed a new method for determining iris liveness. Until now, several approaches have relied on pre-processing, such as iris segmentation, localization, and normalization; however, this method of iris liveness detection is computationally intensive. The suggested method employs Haar transforms on iris images to address this issue, obtaining fragmental coefficients as feature vectors. The Haar transformed iris image fragmental coefficients are used to train various ML and ensemble algorithms. Seven criteria are considered to compare the performance of variants of the suggested approach. Various metrics such as accuracy, precision, recall, f-measure, APCER, BPCER, and ACER are used to check the performance of the models. Presentation attack images are detected with 98.10% accuracy in the Clarkson 2013 dataset. The best accuracy for IIITD-CSD was 97.94%. The experimental results prove the effectiveness of the projected method for detecting iris spoofing attacks. The study’s main contribution is achieving a good accuracy of 98.10% with lesser feature vector size by using the fragmental coefficients of the Haar transformed iris image of size 8 * 8 utilizing a RF algorithm with reduced featured vector size. The reduction in considered feature vector size of iris images with improved accuracy of liveness detection is achieved by exploiting the energy compaction property of Haar transform in the proposed method. The method is tested on three available benchmark datasets for validation of results in a generic form. The cross-dataset validations are performed to prove the worth of the proposed method. The main limitations of this study are as follows: only two pre-processing techniques were applied, such as resizing and converting an image into grayscale. Moreover, this experiment is limited to image size 256 by 256 and only grayscale images are used. In future work, this framework may be extended with the best performance features. Currently, the presented work is limited to the exploration of explored Haar transform features only. However, the hybridization of transform using Haar, DCT and Kekare transforms would be an exciting exploration in the future. Moreover, the proposed framework may be applied for the liveness detection of other biometric traits, like face, fingerprints, etc. The best performance features a level fusion of fragmental coefficients of Haar may be added to this framework in future work.

Author Contributions: Data curation: Smita Khade; Writing original draft: Smita Khade; Supervision: Shilpa Gite, Biswajeet Pradhan; Project administration: Shilpa Gite, Biswajeet Pradhan; Conceptualization: Sudeep Thepade; Methodology: Sudeep Thepade, Shilpa Gite; Validation: Biswajeet Pradhan; Visualization: Sudeep Thepade, Smita Khade, Shilpa Gite, Biswajeet Pradhan; Resources: Biswajeet Pradhan, Abdullah Alamri; Review & Editing: Sudeep Thepade, Biswajeet Pradhan; Funding acquisition: Biswajeet Pradhan, Abdullah Alamri.

Funding Statement: The Centre for Advanced Modelling and Geospatial Information Systems (CAMGIS), Faculty of Engineering and Information Technology, the University of Technology Sydney, Australia, has funded the research. This research is also partially supported by the Researchers Supporting Project No. RSP-2021/14, King Saud University, Riyadh, Saudi Arabia.

Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the present study.

References

  1. Khade, S., Thepade, S. D., Ambedkar, A. (2018). Fingerprint liveness detection using directional ridge frequency with machine learning classifiers. 2018 Fourth International Conference on Computing Communication Control and Automation (ICCUBEA), pp. 1–5. Pune, India.
  2. Khade, S., Thepade, S. D. (2019). Fingerprint liveness detection with machine learning classifiers using feature level fusion of spatial and transform domain features. 2019 5th International Conference On Computing, Communication, Control and Automation (ICCUBEA), pp. 1–6. Pune, India.
  3. Khade, S., Ahirrao, S., Phansalkar, S., Kotecha, K., & Gite, S. (2021). Iris liveness detection for biometric authentication: A systematic literature review and future directions. Inventions, 6(4), 65. [Google Scholar] [CrossRef]
  4. Su, L., & Shimahara, T. (2019). Advanced iris recognition using fusion techniques. NEC Technical Journal, 13(2), 74-77. [Google Scholar]
  5. Kaur, B., Singh, S., & Kumar, J. (2019). Cross-sensor iris spoofing detection using orthogonal features. Computers & Electrical Engineering, 73(1), 279-288. [Google Scholar] [CrossRef]
  6. Khade, S., Ahirrao, S., Thepade, S. (2020). Bibliometric survey on biometric iris liveness detection. Library Philosophy and Practice, 1–29.
  7. Nguyen, K., Fookes, C., Jillela, R., Sridharan, S., & Ross, A. (2017). Long range iris recognition: A survey. Pattern Recognition, 72(9), 123-143. [Google Scholar] [CrossRef]
  8. Khade, S., Gite, S., Thepade, S. D., Pradhan, B., & Alamri, A. (2021). Detection of iris presentation attacks using feature fusion of Thepade’s sorted block truncation coding with gray-level co-occurrence matrix features. Sensors, 21(21), 7408. [Google Scholar] [CrossRef]
  9. Kaur, J., & Jindal, N. (2019). A secure image encryption algorithm based on fractional transforms and scrambling in combination with multimodal biometric keys. Multimedia Tools and Applications, 78(9), 11585-11606. [Google Scholar] [CrossRef]
  10. Choudhary, M., Tiwari, V., & Venkanna, U. (2019). An approach for iris contact lens detection and classification using ensemble of customized DenseNet and SVM. Future Generation Computer Systems, 101(1), 1259-1270. [Google Scholar] [CrossRef]
  11. Chen, Y., Zhang, W. (2018). Iris liveness detection: A survey. 2018 IEEE Fourth International Conference on Multimedia Big Data (BigMM), pp. 1–7. China.
  12. Trokielewicz, M., Czajka, A., Maciejewicz, P. (2016). Human iris recognition in post-mortem subjects: Study and database. 2016 IEEE 8th International Conference on Biometrics Theory, Applications and Systems (BTAS), pp. 1–6. USA.
  13. Fathy, W. S. A., & Ali, H. S. (2018). Entropy with local binary patterns for efficient iris liveness detection. Wireless Personal Communications, 102(3), 2331-2344. [Google Scholar] [CrossRef]
  14. Gupta, R., & Sehgal, P. (2016). A survey of attacks on iris biometric systems. International Journal of Biometrics, 8(2), 145-178. [Google Scholar] [CrossRef]
  15. Khade, S., Thepade, S. D. (2018). Novel fingerprint liveness detection with fractional energy of cosine transformed fingerprint images and machine learning classifiers. 2018 IEEE Punecon, pp. 1–7. Pune, India.
  16. Agarwal, R., Jalal, A. S., & Arya, K. V. (2020). A multimodal liveness detection using statistical texture features and spatial analysis. Multimedia Tools and Applications, 79(19), 13621-13645. [Google Scholar] [CrossRef]
  17. Thavalengal, S., Nedelcu, T., Bigioi, P., & Corcoran, P. (2016). Iris liveness detection for next generation smartphones. IEEE Transactions on Consumer Electronics, 62(2), 95-102. [Google Scholar] [CrossRef]
  18. Hu, Y., Sirlantzis, K., & Howells, G. (2016). Iris liveness detection using regional features. Pattern Recognition Letters, 82, 242-250. [Google Scholar] [CrossRef]
  19. Czajka, A. (2015). Pupil dynamics for iris liveness detection. IEEE Transactions on Information Forensics and Security, 10(4), 726-735. [Google Scholar] [CrossRef]
  20. Fang, M., Damer, N., Boutros, F., Kirchbuchner, F., & Kuijper, A. (2022). The overlapping effect and fusion protocols of data augmentation techniques in iris PAD. Machine Vision and Applications, 33(1), 1-21. [Google Scholar] [CrossRef]
  21. Li, Y. H., Aslam, M. S., Harfiya, L. N., & Chang, C. C. (2021). Conditional wasserstein generative adversarial networks for rebalancing iris image datasets. IEICE Transactions on Information and Systems, 104(9), 1450-1458. [Google Scholar]
  22. Naqvi, R. A., Lee, S. W., Loh, W. K. (2020). Ocular-net: Lite-residual encoder decoder network for accurate ocular regions segmentation in various sensor images. 2020 IEEE International Conference on Big Data and Smart Computing (BigComp), pp. 121–124. Busan, Korea.
  23. Kimura, G. Y., Lucio, D. R., Britto Jr, A. S., Menotti, D. (2020). CNN hyperparameter tuning applied to iris liveness detection. arXiv preprint arXiv:2003.00833.
  24. Khade, S., Gite, S., & Pradhan, B. (2022). Iris liveness detection using multiple deep convolution networks. Big Data and Cognitive Computing, 6(2), 67. [Google Scholar] [CrossRef]
  25. Khade, S., Gite, S., Thepade, S. D., Pradhan, B., & Alamri, A. (2021). Detection of iris presentation attacks using hybridization of discrete cosine transform and haar transform with machine learning classifiers and ensembles. IEEE Access, 9, 169231-169249. [Google Scholar] [CrossRef]
  26. Vyas, R., Kanumuri, T., Sheoran, G., & Dubey, P. (2019). Recent trends of ROI segmentation in iris biometrics: A survey. International Journal of Biometrics, 11(3), 274-307. [Google Scholar] [CrossRef]
  27. Thepade, S. D., Mhaske, V. (2015). New clustering algorithm for vector quantization using hybrid Haar slant error vector. 2015 International Conference on Computing Communication Control and Automation, pp. 634–640. USA.
  28. Yambay, D., Doyle, J. S., Bowyer, K. W., Czajka, A., Schuckers, S. (2014). LivDet-iris 2013-Iris liveness detection competition 2013. IEEE International Joint Conference on Biometrics, pp. 1--8. USA.
  29. Yambay, D., Becker, B., Kohli, N., Yadav, D., Czajka, A., (2017). LivDet-iris 2015 – Iris liveness detection competition 2017. 2017 IEEE International Joint Conference on Biometrics (IJCB), pp. 733–741.
  30. Kohli, N., Yadav, D., Vatsa, M., Singh, R., Noore, A. (2016). Detecting medley of iris spoofing attacks using DESIST. 2016 IEEE 8th International Conference on Biometrics Theory, Applications and Systems (BTAS), pp. 1–6. USA.
  31. Gupta, P., Behera, S., Vatsa, M., Singh, R. (2014). On iris spoofing using print attack. 2014 22nd International Conference on Pattern Recognition, pp. 1681–1686. USA.
  32. Yadav, D., Kohli, N., Doyle, J. S., Singh, R., & Vatsa, M. (2014). Unraveling the effect of textured contact lenses on iris recognition. IEEE Transactions on Information Forensics and Security, 9(5), 851-862. [Google Scholar] [CrossRef]
  33. Fang, M., Damer, N., Boutros, F., Kirchbuchner, F., & Kuijper, A. (2021). Cross-database and cross-attack iris presentation attack detection using micro stripes analyses. Image and Vision Computing, 105(12), 104057. [Google Scholar] [CrossRef]
  34. Arora, S., Bhatia, M. P. S., Kukreja, H. (2020). A multimodal biometric system for secure user identification based on deep learning. International Congress on Information and Communication Technology, pp. 95–103. Singapore.
  35. Omran, M., & AlShemmary, E. N. (2020). An iris recognition system using deep convolutional neural network. Journal of Physics: Conference Series, 1530(1), 012159. [Google Scholar]
  36. Das, P., McFiratht, J., Fang, Z., Boyd, A., Jang, G. et al. (2020). Iris liveness detection competition (livdet-iris)-the 2020 edition. 2020 IEEE International Joint Conference on Biometrics (IJCB), pp. 1–9. Houston, TX.
  37. Wang, K., & Kumar, A. (2019). Cross-spectral iris recognition using CNN and supervised discrete hashing. Pattern Recognition, 86(2), 85-98. [Google Scholar] [CrossRef]
  38. Cheng, Y., Liu, Y., Zhu, X., & Li, S. (2019). A multiclassification method for iris data based on the hadamard error correction output code and a convolutional network. IEEE Access, 7, 145235-145245. [Google Scholar] [CrossRef]
  39. Chatterjee, P., Yalchin, A., Shelton, J., Roy, K., Yuan, X. et al. (2019). Presentation attack detection using wavelet transform and deep residual neural net. International Conference on Security, Privacy and Anonymity in Computation, Communication and Storage, pp. 86–94. Cham.

Cite This Article

Khade, S., Gite, S., Thepade, S. D., Pradhan, B., Alamri, A. (2023). Iris Liveness Detection Using Fragmental Energy of Haar Transformed Iris Images Using Ensemble of Machine Learning Classifiers. CMES-Computer Modeling in Engineering & Sciences, 136(1), 323–345.


cc This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 1804

    View

  • 434

    Download

  • 0

    Like

Share Link