[BACK]
Intelligent Automation & Soft Computing
DOI:10.32604/iasc.2023.025766
images
Article

Metaheuristics with Optimal Deep Transfer Learning Based Copy-Move Forgery Detection Technique

C. D. Prem Kumar1,* and S. Saravana Sundaram2

1Department of Information Technology, Hindusthan College of Engineering and Technology, Coimbatore, 641032, Tamilnadu, India
2Department of Electrical and Electronics Engineering, Hindusthan College of Engineering and Technology, Coimbatore, 641032, Tamilnadu, India
*Corresponding Author: C. D. Prem Kumar. Email: premk.research@gmail.com
Received: 03 December 2021; Accepted: 23 January 2022

Abstract: The extensive availability of advanced digital image technologies and image editing tools has simplified the way of manipulating the image content. An effective technique for tampering the identification is the copy-move forgery. Conventional image processing techniques generally search for the patterns linked to the fake content and restrict the usage in massive data classification. Contrastingly, deep learning (DL) models have demonstrated significant performance over the other statistical techniques. With this motivation, this paper presents an Optimal Deep Transfer Learning based Copy Move Forgery Detection (ODTL-CMFD) technique. The presented ODTL-CMFD technique aims to derive a DL model for the classification of target images into the original and the forged/tampered, and then localize the copy moved regions. To perform the feature extraction process, the political optimizer (PO) with Mobile Networks (MobileNet) model has been derived for generating a set of useful vectors. Finally, an enhanced bird swarm algorithm (EBSA) with least square support vector machine (LS-SVM) model has been employed for classifying the digital images into the original or the forged ones. The utilization of the EBSA algorithm helps to properly modify the parameters contained in the Multiclass Support Vector Machine (MSVM) technique and thereby enhance the classification performance. For ensuring the enhanced performance of the ODTL-CMFD technique, a series of simulations have been performed against the benchmark MICC-F220, MICC-F2000, and MICC-F600 datasets. The experimental results have demonstrated the improvised performance of the ODTL-CMFD approach over the other techniques in terms of several evaluation measures.

Keywords: Copy move detection; image forgery; deep learning; machine learning; parameter tuning; forensics

1  Introduction

With the advancement in the imaging techniques, digital images have become a concrete data source. In the meantime, a considerable amount of image editing mechanisms have placed the authenticity of the images in danger [1]. An objective behindhand image forgery has implemented the manipulation in such a way that it appears to be difficult to view with a naked eye and apply this creation for the malware purposes [2]. However, the forensic investigation has identified the tigers termed as the “paper tigers” [3]. In the same way, in 2008, official images of four Iranian ballistic weapons were proved to be forged as one of the missiles was exposed to be counterfeited [4]. Hence, methods that could guarantee the integrity of images particularly from the evidence centred application are thus required. Recently, the digital image forensics sector has emerged as an interesting field of research that defines the evidence of forgery from that of the digital image [5]. The main objective of digital image forensics is to examine the existence of forgeries in the concerned images by means of employing the passive or active (blind) approaches. The active methodologies such as the digital signatures and the watermarking strategies depend on the data embedded a priori in the image. But, the inaccessibility of the data might restrict the use of active approaches in real life [6]. Therefore, the passive methodologies have been utilized for authenticating the images that don’t need any previous knowledge about them [7]. Image forgery technique consists of image re-sampling, Copy Move forgery, and Image splicing methods respectively [8]. The only objective of the copy-move forgery is to bring further details in the image that wasn’t present initially. This forgery approach can be employed for covering the new data of the images by pasting the copied parts over it. This processing in the image changes the message reflected by the image [9]. Mostly, these transformed images would be applied in the courts of law as an evidence for proving the innocent as guilty. The recognition of copy move image forgery is not an easier task as the copied part is from the original image, hence the features such as; texture, colour, and noise patterns appears to be suitable for the remaining images. Along with this, some post-processing operation could also be employed to the copied segment before pasting it to original image that would make the forgery detection difficult. Different approaches for the copy-move forgery detection have been explained in this work [10].

This study has developed an optimal deep transfer learning based copy move forgery detection (ODTL-CMFD) technique for classifying the target images into the original and the forged/tampered followed by the localization of the copy moved regions. Besides, the ODTL-CMFD technique involves the design of the political optimizer (PO) with the Mobile Networks (MobileNet) model for generating a set of useful vectors. Moreover, an enhanced bird swarm algorithm (EBSA) with least square support vector machine (LS-SVM) model has been applied to the classification process. The parameters (kernel and penalty variables) involved in the LS-SVM model can be adjusted by the use of EBSA. The performance validation of the ODTL-CMFD technique has been carried out using the benchmark MICC-F220, MICC-F2000, and MICC-F600 datasets.

2  Literature Survey

Huang et al. [11], proposed a key point-based image forensics method based on the super pixel segmentation and the Helmert transformation algorithms. The primary objective was to identify the CMFD images and attain the forensic data. Initially, the descriptors and the key points were extracted by utilizing the SIFT approach. Next, based on the descriptors, the matching pair was attained by evaluating the similarities among the key points. Then, this matching pair was grouped based on the spatial distance and the geometric constraint through the Helmert transformation for attaining the coarse forgery region. Later, the isolated areas or the mistakes were removed and the coarse forgery region was well refined. In Meena et al. [12], a strong copy-move image forgery approach with that of the Gaussian-Hermite Moment (GHM) has been proposed. The presented strategy splits the input images into over-lapping blocks of fixed sizes and removes the GHM from the entire set of blocks. A similar block matching can be performed by arranging the individual features lexicographically. In Jindal [13], Image forgery was localized and detected by utilizing the semantic segmentation and the Deep Convolutional Neural Network (DCNN) system. Color illumination was employed for colouring the map after the completion of the pre-processing stage. For training VGG-16, the DCNN-Transfer Learning (TL) method was employed. This technique categorizes the image pixels with that of the forgeries. This categorized image with the colour pixel labels were trained by utilizing the semantic segmentation for limiting the forged pixels. Goel et al. [14], proposed a DL based passive CMFD method which employs a Convolutional Neural Network (CNN) for categorizing the images as forged and original. The CNN method extracts the multi scale features using the distinct kernel sizes. Then, the Combination of the extracted multi scale features was implemented for achieving better recall, accuracy, and precision scores. Abbas et al. [15] aims at utilizing the two advanced DL methods; MobileNetV2 and Smaller VGGNet (inspired from VGGNet). Both these methods appear as time and Eco-friendly DL architectures for detecting the digital image forgeries on the embedded devices. After the completion of a comprehensive review, a modified form of MobileNetV2 that appears to be very efficient on the CMFD that caters for the inconsistencies performed post-forgery was considered. Elaskily et al. [16] examined a novel technique for the CMFD based techniques mainly on the DL approach. The presented method was based on the utilization of the Convolutional Long Short Term Memory (CovLSTM) and the CNN systems. This approach extracts the image features by a sequential number of Convolutional (CNV), ConvLSTM, pooling layers, and matching features. The use of deep learning-based extractors instead of the traditionally handmade ones were considered as an alternative to the more traditional approach. Conventional forensic detectors, on the other hand, are often not real-world accurate in a number of ways, such as the feature extraction strength and the answer to tampering location.

3  The Proposed System

In this study, an effective ODTL-CMFD technique has been developed for the detection and classification of digital images. The ODTL-CMFD technique intends to accomplish the classification of the target images into the original and the forged/tampered using different sub processes such as the MobileNet based feature extraction, the PO based hyper parameter tuning, the LS-SVM based classification, and the EBSA based parameter optimized. Fig. 1 shows the flow of the proposed system.

images

Figure 1: Flow of the proposed system

The detailed functioning process of these modules have been elaborated in the following sections.

3.1 Feature Extraction Using PO with the MobileNet Model

During the feature extraction procedure, the MobileNet model would be applied for deriving a collection of the feature vectors. The CNN generally comprises of the convolution, pooling, and the fully connected (FC) layers [17]. Primarily, the features would be removed by more than one convolutional and a pooling layer. Afterwards, the individual feature maps in the ending convolutional layer would be changed from the 1D vector to the FC. Eventually, the resultant layer would classify the n input images. The network would then alter the weight parameter as the BP and minimize the square variance amongst the classifier outcomes and the predictable outcomes. The neuron in all the layers have been ordered in the 3 dimensional formats: width, height, and depth, whereas the width and the height represent the size of the neurons, and the depth indicates the channel amount of the input picture or the amount of the input feature maps. The framework dependent upon the convolution and the pooling layers has enhanced the robustness of the network model. The CNN has been obtained deeper with that of the multilayer convolutional. With the amount of layers improving, the features attained with learning have been further developed globally. The global feature map that has been learned at the end can be converted to vectors for linking the FC layer. The MobileNet comprises of a lesser framework, minimum computation, and superior precision that can be utilized at the mobile terminals and the embedding devices [18]. According to the depth wise separable convolutional, MobileNets utilize 2 global hyper parameters for maintaining a balance between the efficacy and the accuracy levels. Essential knowledge of the MobileNet is decomposed of convolutional kernels. By means of utilizing the depth wise separable convolutional, the typical convolutional has been observed to be decomposed as the depth wise convolutional and the point wise convolutional with a 1 × 1 convolutional kernel. The depth wise convolutional filter has been found to carry out the convolutional for all the channels, and a 1 × 1 convolutional has been utilized for combining the outcomes of the depth wise convolutional layers. During this manner, N typical convolutional kernels have been modified as the M depth wise convolutional kernels and the N pointwise convolutional kernels. The typical convolution filter joins the input with that of a novel group of outputs, but the depthwise separable convolutional separates an input to 2 layers, one appears to be the filter and the other would be merged suitably.

In addition, the hyper parameters of the MobileNet model can be optimally adjusted by the use of the PO. The PO algorithm is a metaheuristic optimization approach presented in 2020 and is dependent on the multiphase political approach [19]. This PO method stimulates the performance of the politician for achieving the ultimate objective of the optimized. PO implements five sequences of stages for enhancing the provided problem. Similar to the other metaheuristic approaches, the process can be initiated by the population initialization attribute P of size NP. Every row under the population comprises of the number of constituencies and the political parties. The number of input parameters of the problem has been represented by D.

P={Pi,1,Pi,2Pi,D},i=1,2,NP (1)

Furthermore, the party member appear to be a political solution that acts as a selective candidate:

C={ C1,C2,Cn)Cj={Pij,P2j,Pnj} (2)

As previously mentioned, “n” amount of constituencies have been assumed here, where “n” represents the total number of parties competing from all the constituencies C. The optimal member of the party-based fitness has been assumed to be the leader of the party. The process involved in the selection of the party leader has been demonstrated below:

q=armin f(Pij),i{1,2,3,n}and1jn (3)

Pi=Piq

In which P={P1,P2,,Pn} denotes the group of the individual political party leaders. Where Ci indicates the constituency i won by the candidate, the group of the winning candidate’s procedure, the parliamentarian C, has been represented below:

C={C1,C2,,Cn} (4)

In the Election Campaigning stage, every candidate can enhance their majority by assuming the following aspects (1) influencing the constituency members by mentioning about the leader of the party and by itself, (2) candidate can learn from the former elections, and (3) by improving the relative analysis with the constituency winner. The campaigning of the candidate can follow these 3 stages and update their location with that of the preceding location, that is, when the fitness of the candidate is enhanced, the location can be upgraded by utilizing Eq. (5); or else, it can be upgraded by utilizing Eq. (6):

pi,kj(t+1)={m+rand(mpi,ki(t)),ifpi,ki(t1)pi,ki(t)morpi,ki(t1)pi,ki(t)mm+(2×rand1)|mpi,kj(t)|,ifpi,kj(t1)mpi,kj(t)orpi,kj(t1)mpi,kj(t)m+(2×rand1)|mpi,kj(t1)|,ifmpi,kj(t1)pi,kj(t)ormpi,kj(t1)pi,kj(t) (5)

pi,ki(t+1)={m+(2×rand1)|mpi,ki(t)|,ifpi,ki(t1)pi,ki(t)morpi,ki(t1)pi,ki(t)mpi,ki(t1)+rand(pi,ki(t)pi,ki(t1)),ifpi,ki(t1)mpi,ki(t)orpi,ki(t1)mpi,ki(t)m+(2×rand1)|mpi,ki(t1)|,ifmpi,ki(t1)pi,ki(t)ormpi,ki(t1)pi,ki(t) (6)

In the party switching stage, a member of the party Pij can be elected with respect to the adoptive variable λ and can be swapped to another party Pr possessing the minimum fit member Prq . The index q can be calculated by using the following relationship,

q= argmax f(prj),1jn (7)

At the election stage, a candidate is supposed to selectively win according to the fitness as follows

q= argmin f(prj),1jn (8)

cj=pqj (9)

Now, the candidate can be stated as the winner of the constituency by means of upgrading the party leader. During the parliament affairs phase, the government can be formed after the election process. The leader of the parliamentarians and that of the party appears to be categorical; every parliamentarian can upgrade their location by electing the other parliamentarians when there is an enhancement in the fitness level.

3.2 Copy Move Detection Using the Optimal LS-SVM Model

At the time of the copy move detection process, the LS-SVM technique had been used for classifying the target images into the original and the forged/tampered ones. The LSSVM is a well-arranged ML approach dependent upon the statistical learning methodology presented by Vapnik [20]. The LSSVM resolves the high dimension non-linear and the local minimal issues effectively. It can be developed by the SVM technique with 2 additional features. Primarily, the inequality constraint by the equality constraint and secondly it transfers the 2 programming issues to the linear formulas directly. These features have been found to speed up the calculation time of the LSSVM on the SVM technique. The learning procedure of the LSSVM has been illustrated below. The group of instances data sets has been represented as: (xiyi), i = 1, 2, …, n, where xi refers to the ith predictor variables and the yiith represents the outcome variable. Eq. (10) illustrates the linear regression function.

f(x)=wTg(x)+b (10)

where w, b defines the weight vector as well as the deviation correspondingly.

The optimized regression function can be attained by utilizing Eq. (11).

min J(ω,ζ)=12wTw+12Ci=1mζ2 (11)

where C, ζi signifies the error penalty as well as the slack variable correspondingly. The kernel purpose chosen appears to be dependent upon the Mercer’s condition:

K(xi,xj)=ϕ(xi)Tϕ(xj) (12)

The last LSSVM technique has been computed using Eq. (13).

f(x)=i=1maiK(x,xi)+b (13)

where ai implies the Lagrangian multiplier.

The fundamental tuning parameter of the LSSVM technique appears to be the kernel parameter and the penalty C. This analysis possesses the Radial Basis Function (RBF) as the kernel function for training the method. The parameters (kernel and penalty variables) involved in the LS-SVM model can be adjusted by the use of the EBSA. BSA [20] is an effective optimized technique with features of easy procedure, optimum extensibility, and so on. Here N virtual bird flies and forage to food has been assumed here. Supposing xit(i[1,2,,N]) reveal the place of ith bird at t. The bird performances can be explained as:

Foraging performance has been defined as:

xi,jt+1j=xi,jt+(pi,jxi,jt)×C×rand(0,1)+(gi,jxi,jt)×S×rand(0,1) (14)

Vigilance performance has been determined as:

xi,jt+1=xi,jt+A1(meanjxi,jt)×rand(0,1)+A2(pi,jxi,jt)×rand(1,1) (15)

where, A1 and A2 are defined mathematically as:

A1=a1× exp (pFitisumFit+ϵ×N) (16)

A2=a2× exp ((pFitlpFitk|pFitkpFitl|+ϵ)×N×pFitksumFit+ϵ) (17)

a1 and a2 can be the constants from zero and two. ɛ refers to the smaller constant. Flight performance has been defined as:

xi,jt+1=xi,jt+randn(0,1)×xi,jt (18)

xi,jt+1=xi,jt+(xk,jtxi,jt)×FL×randn(0,1) (19)

where FL is in 0 and 2. The chaotic method is a property of sensitivity pertaining to the primary condition. The chaotic signal created by the deterministic methods appears to be the quality of the genus-arbitrariness. Its curve has been defined as the primary value and the chaos mapping parameters. The logistic mapping was utilized extremely practically. The logistic chaotic method is a complex dynamical performance, it could be explained as a variance as in Eq. (20).

λi+1=μ×λi×(1λi) (20)

λ ∈ [0, 1], i = 0, 1, 2, …, μ is in 1 and 4. According to the studies, μ implies the nearby 4, λ refers to the adjacent of the average distribution amongst the zero and the one. In the meantime, the method was entirely chaotic as μ is in 4. A primary population appears to be a vital part of the intelligently optimized technique that controls the convergence rate and the last solution quality [21]. The logistic chaotic mapping was utilized for initializing the population that generates the complete utilization of the data solution space for the purpose of improving the effectiveness of the technique.

The EBSA approach develops a Fitness Function (FF) for obtaining the enhanced classifier efficiency. It defines a positive integer for representing the optimum efficiency of the candidate solution. During this case, the minimal value of the classification error rate assumes the FF, as provided in Eq. (21). An optimum solution appears to be a lesser error rate and the worst solution gains a higher error rate.

fitness(xi)=ClassifierErrorRate(xi)=numberofmisclassifiedi,agesTotalnumberofimages100 (21)

4  Performance Validation

In this section, the simulation analysis of the ODTL-CMFD technique has been accomplished using the Python 3.6.5 tool. Besides, the results have been inspected under three benchmark datasets such as the MICC_F220, the MICC_F2000, and the MICC_F600 datasets [22]. Some of the sample test images have been displayed in Fig. 2.

images

Figure 2: Sample images

Fig. 3 illustrates the sample visualization result analysis of the ODTL-CMFD technique. The first column images portray the original input images and the forged images have been depicted in the second column. Then, the localized forged regions have been showcased in the third column. These figures report the effectiveness of the ODTL-CMFD technique on the detection of the forged regions in the manipulated image.

images

Figure 3: Sample results. a) Original Image b) Forgery Image c) Localization Image

Tab. 1 offers a brief copy move detection performance analysis of the ODTL-CMFD technique under distinct epochs on the MICC_F220 dataset. The experimental values thus point out that the ODTL-CMFD technique has reached the maximum detection performance level under all epochs. For instance, in 10 epochs, the ODTL-CMFD technique has provided accuracy, True positive Rate (TPR_, False Positive Rate (FPR), False Negative Rate (FNR), and True Negative Rate (TNR) of 0.973, 0.955, 0, 0.045, and 1.000. Eventually, with 50 epochs, the ODTL-CMFD technique has attained accuracy, TPR, FPR, FNR, and TNR of 1.000, 0.977, 0, 0.023, and 1.000 respectively.

images

The performance analysis of the ODTL-CMFD approach in terms of the training and the validation accuracy under dissimilar epochs on the MICC_F220 dataset has been illustrated in Fig. 4. The figure illustrates the improved performance of the ODTL-CMFD technique with higher values of training and validation accuracy. It is also ensured that the ODTL-CMFD technique has resulted in increased values of validation accuracy compared with that of the training accuracy.

images

Figure 4: Accuracy analysis of ODTL-CMFD approach on MICC_F220 dataset

Next, the result analysis of the ODTL-CMFD technique with respect to the training and the validation loss under distinct epochs on the MICC_F220 dataset has been illustrated in Fig. 5. The results thus demonstrate that the ODTL-CMFD technique has reached the least training and the validation loss.

images

Figure 5: Loss analysis of the ODTL-CMFD approach on the MICC_F220 dataset

Here, an extensive Receiver Operating Characteristics (ROC) analysis of the ODTL-CMFD approach on the MICC_F220 dataset has been portrayed in Fig. 6. The results demonstrate the enhanced classification performance of the ODTL-CMFD approach with the higher ROC of 99.9578 on the test MICC_F220 dataset. Tab. 2 provides a comparative copy move detection performance analysis of the ODTL-CMFD method under varying epochs on the MICC_F2000 dataset. The experimental values thus point out that the ODTL-CMFD approach has attained the maximal detection performance level under all epochs. For instance, in 10 epochs, the ODTL-CMFD methodology has provided accuracy, TPR, FPR, FNR, and TNR of 0.953, 0.938, 0.067, 0.062, and 0.933. Finally, with 50 epochs, the ODTL-CMFD system has attained accuracy, TPR, FPR, FNR, and TNR of 1.000, 1.000, 0, 0, and 0.972 correspondingly.

images

Figure 6: ROC analysis of the ODTL-CMFD approach on the MICC_F220 dataset

images

The performance analysis of the ODTL-CMFD method with respect to the training and the validation accuracy under dissimilar epochs on the MICC_F2000 dataset has been illustrated in Fig. 7. The figure portrays the improved performance of the ODTL-CMFD method with superior values of training and validation accuracy. It can be also stated that the ODTL-CMFD algorithm has resulted in improved values of validation accuracy related to that of the training accuracy. Afterwards, the outcome analysis of the ODTL-CMFD approach in terms of the training and the validation loss under different epochs on the MICC_F2000 dataset has been illustrated in Fig. 8. The results thus demonstrate that the ODTL-CMFD approach has obtained minimum training and validation loss. An extensive ROC analysis of the ODTL-CMFD method on the MICC_F2000 dataset has been displayed in Fig. 9. The outcomes exhibit the improved classification performance level of the ODTL-CMFD approach with the superior ROC of 99.9731 on the test MICC_F2000 dataset.

images

Figure 7: Accuracy analysis of the ODTL-CMFD approach on the MICC_F2000 dataset

images

Figure 8: Loss analysis of the ODTL-CMFD approach on the MICC_F2000 dataset

images

Figure 9: ROC analysis of the ODTL-CMFD approach on the MICC_F2000 dataset

Tab. 3 suggests a detailed copy move detection performance analysis of the ODTL-CMFD approach under various epochs on the MICC_F600 dataset. The experimental values state that the ODTL-CMFD methodology has achieved a higher detection performance level under all epochs. For instance, in 10 epochs, the ODTL-CMFD approach has offered accuracy, TPR, FPR, FNR, and TNR of 0.946, 0.952, 0.031, 0.049, and 0.969. At last, with 50 epochs, the ODTL-CMFD system has gained accuracy, TPR, FPR, FNR, and TNR of 1.000, 1.000, 0, 0, and 1.000 respectively.

images

The performance analysis of the ODTL-CMFD system with respect to the training and the validation accuracy under dissimilar epochs on the MICC_F600 dataset has been illustrated in Fig. 10. The figure depicts the increased performance level of the ODTL-CMFD algorithm with higher values of training and validation accuracy. It is also to make sure that the ODTL-CMFD approach has resulted in maximum values of validation accuracy against that of the training accuracy. Then, the outcome analysis of the ODTL-CMFD approach with respect to the training and the validation loss under varying epochs on the MICC_F600 dataset has been presented in Fig. 11. The results thus portray that the ODTL-CMFD method has attained minimal training and validation loss.

images

Figure 10: Accuracy analysis of the ODTL-CMFD approach on the MICC_F600 dataset

images

Figure 11: Loss analysis of the ODTL-CMFD approach on the MICC_F600 dataset

At this time, a wide ROC analysis of the ODTL-CMFD approach on the MICC_F600 dataset has been portrayed in Fig. 12. The outcomes display the improved classification performance level of the ODTL-CMFD methodology with the superior ROC of 99.9979 on the test MICC_F600 dataset.

images

Figure 12: ROC analysis of the ODTL-CMFD approach on the MICC_F600 dataset

Finally, the detailed comparative detection performance of the ODTL-CMFD approach takes place using four datasets as in Tab. 4 [2325]. The experimental values thus portray that the Scale Invariant Feature Transform Technique (SIFTT)-CMFD and SIFTT approaches exhibit least copy move detection performance level whereas the SIFTT approach has reached the moderate classification performance level. However, the ODTL-CMFD technique has resulted in the maximum performance level over the other methods on the test datasets. For instance, with the MICC_F220 dataset, the ODTL-CMFD technique has resulted in the TPR of 100%, FPR of 0%, FNR of 0%, and TNR of 100%. Similarly, on the MICC_F2000 dataset, the ODTL-CMFD technique has gained effective outcomes with the TPR of 100%, FPR of 0%, FNR of 0%, and TNR of 100%. Likewise, the ODTL-CMFD technique has accomplished the maximum performance level on the MICC_F600 dataset.

images

Fig. 13 provides a clear CT analysis of the ODTL-CMFD technique with the recent methodologies on the test MICC_F220 dataset. The results show that the SIFTT technique has required a higher CT of 24.13 min. At the same time, the SIFTT-CMFD technique has demonstrated a slightly decreased CT of 17.05 min whereas even the reduced CT of 2.48 min has been desired by the CMFDT technique. However, the ODTL-CMFD technique has offered the least CT of 0.187 min. Fig. 14 exhibits a comparative CT analysis of the ODTL-CMFD technique with the recent methods on the test MICC_F2000 dataset. The experimental values thus report that the SIFTT technique has depicted an increased CT of 312.18 min. Moreover, the SIFTT-CMFD technique has established a somewhat reduced CT of 180.15 min whereas even the reduced CT of 46.58 min has been desired by the CMFDT technique. But the ODTL-CMFD technique has surpassed the other techniques with a minimum CT of 1.201 min.

images

Figure 13: CT analysis of the ODTL-CMFD technique on the MICC_F220 dataset

images

Figure 14: CT analysis of the ODTL-CMFD technique on the MICC_F2000 dataset

Fig. 15 illustrates a detailed CT analysis of the ODTL-CMFD technique with that of the recent methods on the test MICC_F600 dataset. The results thus reveal that the SIFTT technique has demonstrated a superior CT of 115.00 min. Eventually, the SIFTT-CMFD technique has exhibited a certainly decreased CT of 76.21 min whereas even the reduced CT of 17.37 min has been desired by the CMFDT technique. However, the ODTL-CMFD technique has offered the least CT of 0.303 min.

images

Figure 15: CT analysis of the ODTL-CMFD technique on the MICC_F600 dataset

By looking into the above mentioned results and discussions, it has been confirmed that the ODTL-CMFD technique has the ability to effectively detect and classify the copy move forgery compared with that of the recent methods.

5  Conclusion

This paper has introduced an effective ODTL-CMFD technique for the detection and classification of the digital images. The ODTL-CMFD technique intends to accomplish the classification of the target images into the original and the forged/tampered, followed by the localization of the copy moved regions. In addition, the ODTL-CMFD technique employs the MobileNet based feature extraction and the PO based hyper parameter tuning process. Moreover, the EBSA with the LS-SVM model has been applied for the classification of the images into the actual or the forged images. The performance validation of the ODTL-CMFD technique has been carried out using the benchmark MICC_F220, MICC_F2000, and MICC_F600 datasets. The experimental results thus state the better performance of the ODTL-CMFD technique over the recent approaches with respect to various evaluation measures. Therefore, the ODTL-CMFD technique can be utilized as an effective tool for the copy move forgery detection and classification. In future, the hybrid DL models can be utilized instead of the MobileNet and the LS-SVM models for enhancing the detection performance levels.

Funding Statement: The authors received no specific funding for this study.

Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the present study.

References

  1. J. Zhao and J. Guo, “Passive forensics for copy-move image forgery using a method based on DCT and SVD,” Forensic Science International, vol. 233, no. 1–3, pp. 158–166, 2013.
  2. N. Krawetz, “A pictures worth digital image analysis and forensics,” in Hacker Factor Solutions, Black Hat Briefings, First Edition, Las Vegas, United States of America, 2007.
  3. F. M. Al_Azrak, A. Sedik, M. I. Dessowky, G. M. El Banby, A. A. Khalaf et al., “An efficient method for image forgery detection based on trigonometric transforms and deep learning,” Multimedia Tools and Applications, vol. 79, no. 25, pp. 18221–18243, 2020.
  4. A. Thakur and N. Jindal, “Hybrid deep learning and machine learning approach for passive image forensic,” IET Image Processing, vol. 14, no. 10, pp. 1952–1959, 2020.
  5. S. P. Chalamalasetty and S. R. Giduturi, “Research perception towards copy-move image forgery detection: Challenges and future directions,” International Journal of Image and Graphics, vol. 21, no. 4, pp. 1–18, 2021.
  6. R. Agarwal and O. P. Verma, “An efficient copy move forgery detection using deep learning feature extraction and matching algorithm,” Multimedia Tools and Applications, vol. 79, no. 11, pp. 7355–7376, 2020.
  7. Z. J. Barad and M. M. Goswami, “Image forgery detection using deep learning: A survey,” in Proc. IEEE 6th Int. Conf. on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, pp. 571–576, 2020.
  8. M. M. Islam, G. Karmakar, J. Kamruzzaman and M. Murshed, “A robust forgery detection method for copy–Move and splicing attacks in images,” Electronics, vol. 9, no. 9, pp. 500–1524, 2020.
  9. Y. Zhu, C. Chen, G. Yan, Y. Guo and Y. Dong, “AR-Net: Adaptive attention and residual refinement network for copy-move forgery detection,” IEEE Transactions on Industrial Informatics, vol. 16, no. 10, pp. 6714–6723, 2020.
  10. G. Gani and F. Qadir, “A robust copy-move forgery detection technique based on discrete cosine transform and cellular automata,” Journal of Information Security and Applications, vol. 54, pp. 102510–102524, 2020.
  11. H. Y. Huang and A. J. Ciou, “Copy-move forgery detection for image forensics using the superpixel segmentation and the helmert transformation,” EURASIP Journal on Image and Video Processing, vol. 2019, no. 1, pp. 1–16, 2019.
  12. K. B. Meena and V. Tyagi, “A copy-move image forgery detection technique based on Gaussian-hermite moments,” Multimedia Tools and Applications, vol. 78, no. 23, pp. 33505–33526, 2019.
  13. N. Jindal, “Copy move and splicing forgery detection using deep convolution neural network, and semantic segmentation,” Multimedia Tools and Applications, vol. 80, no. 3, pp. 3571–3599, 2021.
  14. N. Goel, S. Kaur and R. Bala, “Dual branch convolutional neural network for copy move forgery detection,” IET Image Processing, vol. 15, no. 3, pp. 656–665, 2021.
  15. M. N. Abbas, M. S. Ansari, M. N. Asghar, N. Kanwal, T. O’Neill et al., “Lightweight deep learning model for detection of copy-move image forgery with post-processed attacks,” in Proc. IEEE 19th World Symp. on Applied Machine Intelligence and Informatics (SAMI), Herl’any, Slovakia, pp. 125–130, 2021.
  16. M. A. Elaskily, M. H. Alkinani, A. Sedik and M. M. Dessouky, “Deep learning based algorithm (ConvLSTM) for copy move forgery detection,” Journal of Intelligent & Fuzzy Systems, vol. 40, no. 3, pp. 4385–4405, 2021.
  17. X. B. Meng, X. Z. Gao, L. Lu, Y. Liu and H. Zhang, “A new bio-inspired optimization algorithm: Bird swarm algorithm,” Journal of Experimental & Theoretical Artificial Intelligence, vol. 28, no. 4, pp. 673–687, 2015.
  18. W. Wang, Y. Hu, T. Zou, H. Liu, J. Wang et al., “A new image classification approach via improved MobileNet models with local receptive field expansion in shallow layers,” Computational Intelligence and Neuroscience, vol. 2020, no. 8817849, pp. 1–10, 2020.
  19. V. Basetti, S. S. Rangarajan, C. K. Shiva, H. Pulluri and R. Kumar, “Economic emission load dispatch problem with valve-point loading using a novel quasi-oppositional-based political optimizer,” Electronics, vol. 10, no. 21, pp. 2596–2612, 2021.
  20. D. J. Armaghani, D. Kumar, P. Samui, M. Hasanipanah and B. Roy, “A novel approach for forecasting of ground vibrations resulting from blasting: Modified particle swarm optimization coupled extreme learning machine,” Engineering with Computers, vol. 37, no. 4, pp. 3221–3235, 2021.
  21. D. Zhang, J. Yang and P. Yang, “An improved chaos bird swarm optimization algorithm,” Journal of Physics: Conference Series, vol. 1176, no. 2, pp. 1–9, 2019.
  22. I. Amerini, L. Ballan, R. Caldelli, A. D. Bimbo and G. Serra, “A SIFT-based forensic method for copy-move attack detection and transformation recovery,” IEEE Transactions on Information Forensics and Security, vol. 6, no. 3, pp. 1099–1110, 2011.
  23. I. Amerini, L. Ballan, R. Caldelli, A. D. Bimbo, L. D. Tongo et al., “Copy-move forgery detection and localization by means of robust clustering with J-linkage,” Signal Processing: Image Communication, vol. 28, no. 6, pp. 659–669, 2013.
  24. I. Amerini, L. Ballan, R. Caldelli, A. Del Bimbo and G. Serra, “A SIFT-based forensic method for copy–move attack detection and transformation recovery,” IEEE Transactions on Information Forensics and Security, vol. 6, no. 3, pp. 1099–1110, 2011.
  25. M. A. Elaskily, H. A. Elnemr, A. Sedik, M. M. Dessouky, G. M. E. Banby et al., “A novel deep learning framework for copy-moveforgery detection in images,” Multimedia Tools & Applications, vol. 79, pp. 19167–19192, 2020.
images This work is licensed under a Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.