999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看

404 Not Found


nginx
404 Not Found

404 Not Found


nginx
404 Not Found

404 Not Found


nginx
404 Not Found

404 Not Found


nginx
404 Not Found

404 Not Found


nginx
404 Not Found

404 Not Found


nginx

Prediction of Changed Faces with HSCNN

2022-08-24 03:30:56JinhoHan
Computers Materials&Continua 2022年5期

Jinho Han

Department of Liberal Studies(Computer),Korean Bible University,Seoul,01757,Korea

Abstract: Convolutional Neural Networks (CNN) have been successfully employed in the field of image classification.However, CNN trained using images from several years ago may be unable to identify how such images have changed over time.Cross-age face recognition is,therefore,a substantial challenge.Several efforts have been made to resolve facial changes over time utilizing recurrent neural networks(RNN)with CNN.The structure of RNN contains hidden contextual information in a hidden state to transfer a state in the previous step to the next step.This paper proposes a novel model called Hidden State-CNN(HSCNN).This adds to CNN a convolution layer of the hidden state saved as a parameter in the previous step and requires no more computing resources than CNN.The previous CNN-RNN models perform CNN and RNN, separately and then merge the results.Therefore, their systems consume twice the memory resources and CPU time,compared with the HSCNN system,which works the same as CNN only.HSCNN consists of 3 types of models.All models load hidden state ht-1 from parameters of the previous step and save ht as a parameter for the next step.In addition,model-B adds ht-1 to x,which is the previous output.The summation of ht-1 and x is multiplied by weight W.In model-C the convolution layer has two weights:W1 and W2.Training HSCNN with faces of the previous step is for testing faces of the next step in the experiment.That is, HSCNN trained with past facial data is then used to verify future data.It has been found to exhibit 10 percent greater accuracy than traditional CNN with a celeb face database.

Keywords: CNN-RNN;HSCNN;hidden state;changing faces

1 Introduction

Face recognition(FR)systems have been continually developed for personal authentication.These efforts have resulted in FR applications acting on mobile phones [1].Researchers have proposed several ideas for FR systems:eigen faces[2],independent component analysis[3],linear discriminant analysis [4,5], three-dimensional (3D) method [6–9], and liveness detection schemes to prevent the misuse of photographic images[10].Using data acquisition methodology,Jafri et al.[11]divided FR techniques into three categories:intensity images, video sequences, and 3D or infra-red techniques.They introduced AI approaches as one of the operating methods for intensity images and reported that it worked efficiently for somewhat complex FR scenarios.Such techniques had not previously been utilized for practical everyday purposes.

In 2012,AlexNet[12]was proposed and became a turning point in large-scale image recognition.It was the first CNN,one of the deep learning techniques,and was declared the winner of the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) 2012 with 83.6% accuracy.In ILSVRC 2013,Clarifai was the winner with 88.3%[13,14],whereas in ILSVRC 2014,GoogLeNet was the winner with 93.3% [15].The latter was an astonishing result because humans trained for annotator comparison exhibited approximately 95%accuracy in ILSVRC[16].In 2014,using a nine-layer CNN,DeepFace[17]achieved 97.35%accuracy in FR,closely approaching the 97.53%ability of humans to recognize cropped Labeled Faces in the Wild(LFW)benchmark[18].However,DeepID2[19]achieved 99.15%face verification accuracy with the balanced identification and verification features on ConvNet,which contained four convolution layers.In 2015,DeepID3[20]achieved 99.53%accuracy using VGGNet(Visual Geometry Group Net)[21],whereas FaceNet[22]achieved 99.63%using only 128-bytes per face.

CNN consists of convolution layers,pooling layers,and fully connected layers.However,a number of problems still needed to be addressed.For instance,CNN trained with past images failed to verify changed images according to a time sequence.In their in-depth FR survey,Wang et al.[23]described three types of cross-factor FR algorithms as challenges to address in real-world applications:crosspose,cross-age,and makeup.Cross-age FR is a substantial challenge with respect to facial aging over time.Several researchers have attempted to resolve this issue.For instance,Liu et al.[24]proposed an age estimation system for faces with CNN.Bianco et al.[25]and Khiyari et al.[26]applied CNN to learn cross-age information.Li et al.[27]suggested metric learning in a deep CNN.Other studies have suggested combining CNN with recurrent neural networks(RNN)to verify changed images because RNN can predict data sequences [28].RNN contains contextual information in a hidden state to transfer a state in the previous step to the next step, and has been found to generate sequences in various domains,including text[29],motion capture data[30],and music[31,32].

This paper proposes a novel model called Hidden State-CNN (HSCNN) as well as training the modified CNN with past data to verify future data.HSCNN adds to CNN a convolution layer of the hidden state saved as a parameter.The contributions of the present study are as follows:

First,the proposed model,HSCNN,exhibits 10 percent greater accuracy than traditional CNN with a celeb face database[33].Facial images of the future were tested after training based on facial images of the past.HSCNN adds the hidden state saved as a parameter in the previous step to the CNN structure.Further details on this process are provided in Section 4.2.

Secondly, because HSCNN included the hidden state of RNN in the proposed architecture, it was efficient in its use of computing resources.Other researchers have performed CNN and RNN,separately, and merged the results in the system, consuming double the number of resources and processing time.Further details are presented in Section 2.

Thirdly,this paper indicates that HSCNN can train with only two images of one person per step.Also,the loss value reached 0.4 in just 40 epochs in training with loading parameters and 250 epochs in training without loading parameters.Therefore,the HSCNN achieves efficiency because it uses only two images and trains in 40 epochs with loading parameters.This is explained further in Section 4.1.

In the remainder of this paper,Section 2 introduces related works,Section 3 outlines the proposed method,Section 4 presents the experimental results,and Section 5 provides the conclusion.

2 Related Works

Some neural network models can acquire contextual information in various text environments using recurrent layers.Convolutional Recurrent Neural Networks (CRNN) works for a scene text recognition system to read scene texts in the image [34].It contains both convolutional and LSTM recurrent layers of the network architecture and uses past state and current input to predict the subsequent text.Recurrent Convolutional Neural Networks(RCNN)also uses a recurrent structure to classify text from document datasets[35].The combined CNN and RNN model uses relations between phrases and word sequences[36]and in the field of natural language processing(NLP)[37].

Regarding changed images, combining CNN with RNN methods have been proposed in image classification[38]and a medical paper for blood cell images[39].These authors merged the features extracted from CNN and RNN to determine the long-term dependency and continuity relationship.A CNN-LSTM algorithm was proposed for stock price prediction according to leading indicators[40].This algorithm employed a sequence array of historical data as the input image of the CNN and feature vectors extracted from the CNN as the input vector of LSTM.However, these methods used CNN and RNN,separately,and merged the results vectors extracted from CNN to RNN.Therefore,their systems consume twice the memory resources and CPU time, compared with the proposed system,which works the same as CNN only.Fig.1 presents an overview of the models developed by Yin et al.[38]and Liang et al.[39].

Figure 1:Overview of previous CNN-RNN models

Han introduced incremental learning in CNN[41].Incremental-CNN(I-CNN)was tested using the MNIST dataset.HSCNN referenced I-CNN, which used hidden states (ht-1, ht) and an added convolution layer.For training,I-CNN used the MNIST database of handwritten digits comprising 60,000 examples and changed handwritten digits(CHD)comprising 1,000 images.This paper proposes HSCNN, a new structure combining CNN with RNN.It adds a hidden state of RNN into a convolution layer of CNN.Consequently,HSCNN acts like CNN and performs efficiently for crossage FR.

3 Proposed Method:Hidden State CNN

The following subsections explain the equation of the loss function cross-entropy error and optimizer of the stochastic gradient descent used in the proposed method.And then,3 types of models of Hidden State CNN are described.

3.1 Loss Function and Optimizer

In HSCNN, experiments indicated that cross-entropy error (CEE) was the appropriate loss function (cost function or objective function).The best model has a cross-entropy error of 0.The smaller the CEE,the better the model.The CEE equation is:

where tkis the truth label,and ykis the softmax probability for the kthclass.When calculating the log,a minimal value near the 0.0 delta is necessary to prevent a ‘0’error.In the python code, the CEE equation is:

Also,the optimizer is the stochastic gradient descent(SGD),which is a method for optimizing a loss function.The equation is:

where W is the weight,ηis the learning rate,andis the gradient of the loss function L for W.

3.2 Hidden State CNN(HSCNN)

HSCNN consists of 3 types of models:model-A,model-B,and model-C.Model-A is the same as I-CNN.Fig.2 presents the overall structure of HSCNN,which is the modified AlexNet with hidden state htof Convolution layer 5.HSCNN consists of convolution layers,RelU function,pooling layers,affine layers,and soft max function such as traditional CNN,but it also has hidden states:ht-1and ht.This distinguishes HSCNN from CNN.

Figure 2:The overall structure of HSCNN

Fig.3 presents the structure of model-A.Model-A loads hidden state ht-1from the parameters of the previous step and saves htas a parameter for the next step.It adds the convolution layer 6 and W2.

The activation function of the rectified linear unit(RelU)equation of model-A is:

Fig.4 presents the structure of model-B.Model-B loads hidden state ht-1from parameters of the previous step and saves htas a parameter for the next step.The model adds ht-1to x,which was the previous output.The summation of ht-1and x is multiplied by weight W.The activation function of the rectified linear unit(RelU)equation of model-B is:

Figure 3:The structure of model-A of HSCNN

Figure 4:The structure of model-B of HSCNN

Fig.5 presents the structure of model-C.Model-C loads hidden state ht-1from parameters of the previous step and saves ht as av parameter for the next step.In this case,the convolution layer has two weights:W1and W2.The activation function of the rectified linear unit(RelU)equation of model-C is:

Figure 5:The structure of model-C of HSCNN

4 Experimental Results

In the experiment,HSCNN used Cross-Entropy Error as the loss function,Stochastic Gradient Descent(SGD)as the optimizer,and the Instruction rate(Ir)was 0.001.HSCNN also used python,NumPy library,and CuPy library for the NVIDIA GPUs employed for software coding.

4.1 Preparations for Experiments

The essential items required for the experiments are listed in Tab.1.The experiments used FaceApp Pro to make old faces from young faces and Abrosoft Fanta Morph Deluxe to generate 100 morphing facial images between young and old faces.

Table 1:Items prepared for the experiments

The facial images of 1,100 persons were selected from the celeb face database,known as the Largescale CelebFaces Attributes (CelebA) Dataset.Those faces were then paired with a similar face; for example,man to man,woman to woman,Asian to Asian,and so on.One was chosen as the young face between paired faces and the other was changed to the old face.To make these changes,the photo editing application FaceApp Pro was used.Between the transition from young faces to old faces,100 continually changing faces were created using the photo morphing software FantaMorph.

The difference between the young and old images of the same person made by FaceApp Pro was not sufficient, so the experiments also paired the faces of two other persons.Thus, the young face and old face of each pair were different persons.Finally,102 images changing from young to old were created for all 550 paired images.Number 1 is the youngest image,and number 102 is the oldest.102 images show the aging state over time from young to old.Number 1 is the youngest image,and number 102 is the oldest.102 images show the aging state over time from young to old.Numbers 1 to 10 are young images used as the first training data,and the last numbers 101 to 102 are old images and used as final target images.Using 90 images from numbers 11 to 100,the experiments used 10 steps to learn and test aging changes over time.When learning using 9 images in each of 10 steps,all 90 images are used.However,in this case,learning occurs on the entire data,and the training result of the proposed model or CNN was almost the same.So the prediction experiments become meaningless.Therefore,to clearly confirm the results of the prediction experiment,two images are trained for each step,and the two images mean a specific point in the middle of facial aging.This is the same as testing the face at the last oldest point with two final target images.Tab.2 presents sample images from the dataset and their numbers for each step:primary step,target step,and step 1 to step 10.Beyond primary step,each step has only two samples.There is a substantial difference between images number 60 to number 90;however,after number 90,all facial images look almost identical.

Table 2:Dataset samples of changed images according to steps

HSCNN appears to be an extremely efficient method as it achieved 99.9%accuracy and 0.01 loss value with only two training images in each step from step 1 to step 10.Also,the HSCNN is trained with loading parameters saved in the previous step.The use of loading parameters means the training epochs can be shortened.Fig.6 indicates that the loss value reaches 0.4 in 250 epochs in training without loading parameters and just 40 epochs with loading parameters.Thus,the HSCNN achieves efficiency because it uses only two images and trains in 40 epochs with loading parameters.

4.2 Experiments

The experiments utilized AlexNet as the traditional CNN and created three models of HSCNN modifying AlexNet.All models added hidden states to the convolution five-layer and used the RelU 5 layer of the AlexNet.

Figure 6:The efficiency of HSCNN

Fig.7 presents the samples of train steps of the experiments and it shows the accuracy of each training step and loss values through 150 epochs.In the case of training step 3 of model-A,the graph shows the accuracy of training step 3, testing of step 4, step 5, step 6, step 7, step 8, step 9, step 10,and target step.It also shows the loss values according to the epochs.For including loss values in one graph,loss values were divided by 10.

Figure 7:The samples of train steps of the experiments

The experiments consisted of 2 groups:experiment I,experiment II.Experiment I tested the target step,and experiment II tested step 3(behind the training step).Step 1 employed the parameters saved in the primary step,while step 2 used the parameters saved in step 1.

Tab.3 presents the accuracy of the target step in testing faces.Tab.4 presents the accuracy of step 3 behind,according to the models,in testing faces.The test of step 3 behind means that HSCNN tested step 4 when training step 1,and step 5 when training step 2.

Table 3:Accuracy(%)of testing the target step(Experiment I)

Table 4:Accuracy(%)of testing the step 3 behind(Experiment II)

Based on Tabs.3 and 4,the results of the experiments are presented in Figs.8–10.Fig.8 depicts the differences between model-A and AlexNet.These indicate that model-A achieved more than 10%higher accuracy in step 3 to step 5 of experiment I and in step 1 to step 3 of experiment II.Therefore,model-A can be employed for both long-and short-time changes.

Fig.9 depicts the differences between model-B and AlexNet.These indicate that model-B achieved more than 10%higher accuracy in step 4 to step 6 of experiment I and in step 1 to step 2 of experiment II.Therefore,model-B can be employed to predict both long-and short-time changes.Fig.10 depicts the differences between model-C and AlexNet.These indicate that model-C achieved more than 10%higher accuracy in step 2 to step 3 of experiment II, but not in any step of experiment I.Therefore,model-C is considered an appropriate method to predict short-time changes.

Figure 8:The results of the model-A and AlexNet experiments

Figure 9:The results of model-B and AlexNet experiments

Figure 10:The results of model-C and AlexNet experiments

5 Conclusion

This paper proposes a novel model, Hidden State-CNN (HSCNN), which adds to CNN a convolution layer of the hidden state.Training with past data to verify future data was the aim of the experiments,and HSCNN exhibited 10 percent higher accuracy with the celeb face database than the traditional CNN.The experiments consisted of 2 groups.Experiment I tested the target step and experiment II tested step 3 behind the training step.The latter step means that step 4 was tested when training step 1.Experiment I assessed long-time changes, and experiment II relatively shorttime changes.Both yielded similar results.Furthermore, HSCNN achieved a more efficient use of computing resources because its structure differed from that of other CNN-RNN methods.HSCNN also represented a new and efficient training process to verify changing faces.It used only two training images in each step and achieved 99.9%accuracy and 0.01 loss value.

Funding Statement:This work was supported by the National Research Foundation of Korea(NRF)grant in 2019(NRF-2019R1G1A1004773).

Conflicts of Interest:The author declares that he has no conflicts of interest to report regarding the present study.

404 Not Found

404 Not Found


nginx
404 Not Found

404 Not Found


nginx
404 Not Found

404 Not Found


nginx
404 Not Found

404 Not Found


nginx
404 Not Found

404 Not Found


nginx
主站蜘蛛池模板: 国产麻豆va精品视频| 国产亚洲精品自在线| 国产美女在线免费观看| 国产福利免费在线观看| 国产男女XX00免费观看| a级毛片免费看| 美美女高清毛片视频免费观看| 亚洲欧洲日韩综合| 国产乱人乱偷精品视频a人人澡 | 中文字幕有乳无码| 丁香五月婷婷激情基地| 亚洲精品无码AV电影在线播放| 国产自在线播放| 福利在线不卡一区| 国产超碰在线观看| 国产精选自拍| 99热线精品大全在线观看| 亚洲三级视频在线观看| 国产成人亚洲综合a∨婷婷| 波多野结衣AV无码久久一区| 无遮挡一级毛片呦女视频| 亚洲一级无毛片无码在线免费视频| 久久免费观看视频| 免费高清毛片| 精品无码国产一区二区三区AV| 久久久噜噜噜| 国产福利在线观看精品| 亚洲欧美激情小说另类| 中文字幕日韩欧美| 国产欧美高清| 日韩欧美91| 91精品情国产情侣高潮对白蜜| 亚洲中文无码h在线观看| 成人午夜福利视频| 亚洲日韩每日更新| 国产第一色| 欧美在线一级片| 久久香蕉欧美精品| 精品国产成人高清在线| 青青草国产免费国产| 九九免费观看全部免费视频| 久久公开视频| 亚洲91精品视频| 自慰高潮喷白浆在线观看| 欧美精品在线观看视频| 日韩欧美国产区| 青草视频久久| 欧美成人二区| 小13箩利洗澡无码视频免费网站| 久久a级片| 亚洲欧洲日韩久久狠狠爱| 天天干伊人| 亚洲欧美日韩中文字幕在线| 欧美成人精品一级在线观看| 青青青国产精品国产精品美女| 中文字幕在线视频免费| 又爽又大又黄a级毛片在线视频| 亚洲久悠悠色悠在线播放| 无码网站免费观看| 亚洲av综合网| 美女一级毛片无遮挡内谢| 成人韩免费网站| 无码精品福利一区二区三区| 久草网视频在线| 黄片一区二区三区| 国产精品第页| 日本午夜精品一本在线观看| 国产激情国语对白普通话| 国产精品亚洲五月天高清| 18禁色诱爆乳网站| 欧美日韩另类国产| 免费在线a视频| 无码AV日韩一二三区| 亚洲午夜综合网| 国产精品国产三级国产专业不| 欧美日韩精品一区二区在线线| 日本三区视频| 亚洲Av综合日韩精品久久久| 欧美久久网| 色一情一乱一伦一区二区三区小说| 国产91视频观看| 精品国产Av电影无码久久久|