Visible to the public Biblio

Filters: Keyword is image recognition  [Clear All Filters]
2023-07-21
Sadikoğlu, Fahreddin M., Idle Mohamed, Mohamed.  2022.  Facial Expression Recognition Using CNN. 2022 International Conference on Artificial Intelligence in Everything (AIE). :95—99.
Facial is the most dynamic part of the human body that conveys information about emotions. The level of diversity in facial geometry and facial look makes it possible to detect various human expressions. To be able to differentiate among numerous facial expressions of emotion, it is crucial to identify the classes of facial expressions. The methodology used in this article is based on convolutional neural networks (CNN). In this paper Deep Learning CNN is used to examine Alex net architectures. Improvements were achieved by applying the transfer learning approach and modifying the fully connected layer with the Support Vector Machine(SVM) classifier. The system succeeded by achieving satisfactory results on icv-the MEFED dataset. Improved models achieved around 64.29 %of recognition rates for the classification of the selected expressions. The results obtained are acceptable and comparable to the relevant systems in the literature provide ideas a background for further improvements.
Shiomi, Takanori, Nomiya, Hiroki, Hochin, Teruhisa.  2022.  Facial Expression Intensity Estimation Considering Change Characteristic of Facial Feature Values for Each Facial Expression. 2022 23rd ACIS International Summer Virtual Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing (SNPD-Summer). :15—21.
Facial expression intensity, which quantifies the degree of facial expression, has been proposed. It is calculated based on how much facial feature values change compared to an expressionless face. The estimation has two aspects. One is to classify facial expressions, and the other is to estimate their intensity. However, it is difficult to do them at the same time. There- fore, in this work, the estimation of intensity and the classification of expression are separated. We suggest an explicit method and an implicit method. In the explicit one, a classifier determines which types of expression the inputs are, and each regressor determines its intensity. On the other hand, in the implicit one, we give zero values or non-zero values to regressors for each type of facial expression as ground truth, depending on whether or not an input image is the correct facial expression. We evaluated the two methods and, as a result, found that they are effective for facial expression recognition.
2023-07-18
Nguyen, Thanh Tuan, Nguyen, Thanh Phuong, Tran, Thanh-Hai.  2022.  Detecting Reflectional Symmetry of Binary Shapes Based on Generalized R-Transform. 2022 International Conference on Multimedia Analysis and Pattern Recognition (MAPR). :1—6.
Analyzing reflectionally symmetric features inside an image is one of the important processes for recognizing the peculiar appearance of natural and man-made objects, biological patterns, etc. In this work, we will point out an efficient detector of reflectionally symmetric shapes by addressing a class of projection-based signatures that are structured by a generalized \textbackslashmathcalR\_fm-transform model. To this end, we will firstly prove the \textbackslashmathcalR\_fmˆ-transform in accordance with reflectional symmetry detection. Then different corresponding \textbackslashmathcalR\_fm-signatures of binary shapes are evaluated in order to determine which the corresponding exponentiation of the \textbackslashmathcalR\_fm-transform is the best for the detection. Experimental results of detecting on single/compound contour-based shapes have validated that the exponentiation of 10 is the most discriminatory, with over 2.7% better performance on the multiple-axis shapes in comparison with the conventional one. Additionally, the proposed detector also outperforms most of other existing methods. This finding should be recommended for applications in practice.
2023-06-30
Gupta, Rishabh, Singh, Ashutosh Kumar.  2022.  Privacy-Preserving Cloud Data Model based on Differential Approach. 2022 Second International Conference on Power, Control and Computing Technologies (ICPC2T). :1–6.
With the variety of cloud services, the cloud service provider delivers the machine learning service, which is used in many applications, including risk assessment, product recommen-dation, and image recognition. The cloud service provider initiates a protocol for the classification service to enable the data owners to request an evaluation of their data. The owners may not entirely rely on the cloud environment as the third parties manage it. However, protecting data privacy while sharing it is a significant challenge. A novel privacy-preserving model is proposed, which is based on differential privacy and machine learning approaches. The proposed model allows the various data owners for storage, sharing, and utilization in the cloud environment. The experiments are conducted on Blood transfusion service center, Phoneme, and Wilt datasets to lay down the proposed model's efficiency in accuracy, precision, recall, and Fl-score terms. The results exhibit that the proposed model specifies high accuracy, precision, recall, and Fl-score up to 97.72%, 98.04%, 97.72%, and 98.80%, respectively.
2023-03-31
Zhang, Hui, Ding, Jianing, Tan, Jianlong, Gou, Gaopeng, Shi, Junzheng.  2022.  Classification of Mobile Encryption Services Based on Context Feature Enhancement. 2022 IEEE Asia-Pacific Conference on Image Processing, Electronics and Computers (IPEC). :860–866.
Smart phones have become the preferred way for Chinese Internet users currently. The mobile phone traffic is large from the operating system. These traffic is mainly generated by the services. In the context of the universal encryption of the traffic, classification identification of mobile encryption services can effectively reduce the difficulty of analytical difficulty due to mobile terminals and operating system diversity, and can more accurately identify user access targets, and then enhance service quality and network security management. The existing mobile encryption service classification methods have two shortcomings in feature selection: First, the DL model is used as a black box, and the features of large dimensions are not distinguished as input of classification model, which resulting in sharp increase in calculation complexity, and the actual application is limited. Second, the existing feature selection method is insufficient to use the time and space associated information of traffic, resulting in less robustness and low accuracy of the classification. In this paper, we propose a feature enhancement method based on adjacent flow contextual features and evaluate the Apple encryption service traffic collected from the real world. Based on 5 DL classification models, the refined classification accuracy of Apple services is significantly improved. Our work can provide an effective solution for the fine management of mobile encryption services.
Ren, Zuyu, Jiang, Weidong, Zhang, Xinyu.  2022.  Few-Shot HRRP Target Recognition Method Based on Gaussian Deep Belief Network and Model-Agnostic Meta-Learning. 2022 7th International Conference on Signal and Image Processing (ICSIP). :260–264.
In recent years, radar automatic target recognition (RATR) technology based on high-resolution range profile (HRRP) has received extensive attention in various fields. However, insufficient data on non-cooperative targets seriously affects recognition performance of this technique. For HRRP target recognition under few-shot condition, we proposed a novel gaussian deep belief network based on model-agnostic meta-learning (GDBN-MAML). In the proposed method, GDBN allowed real-value data to be transmitted over the entire network, which effectively avoided feature loss due to binarization requirements of conventional deep belief network (DBN) for data. In addition, we optimized the initial parameters of GDBN by multi-task learning based on MAML. In this way, the number of training samples required by the model for new recognition tasks could be reduced. We applied the proposed method to the HRRP recognition experiments of 3 types of 3D simulated aircraft models. The experimental results showed that the proposed method had higher recognition accuracy and generalization performance under few-shot condition compared with conventional deep learning methods.
2022-07-14
Jiang, Qingwei.  2021.  An Image Hiding Algorithm based on Bit Plane and Two-Dimensional Code. 2021 Third International Conference on Intelligent Communication Technologies and Virtual Mobile Networks (ICICV). :851–854.
An image hiding algorithm based on bit plane and two-dimensional code is proposed in this paper. The main characteristic of information hiding is to use the information redundant data of the existing image, to embed the information into these redundant data by the information hiding algorithm, or to partially replace redundant information with information to be embedded to achieve a visual invisible purpose. We first analyze the color index usage frequency of the block index matrix in the algorithm, and calculate the distance between the color of the block index matrix with only one color and the other color in the palette that is closest to the color. Then, the QR model and the compression model are applied to improve the efficiency. We compare the proposed model with the stateof-the-art models.
2022-07-05
Schoneveld, Liam, Othmani, Alice.  2021.  Towards a General Deep Feature Extractor for Facial Expression Recognition. 2021 IEEE International Conference on Image Processing (ICIP). :2339—2342.
The human face conveys a significant amount of information. Through facial expressions, the face is able to communicate numerous sentiments without the need for verbalisation. Visual emotion recognition has been extensively studied. Recently several end-to-end trained deep neural networks have been proposed for this task. However, such models often lack generalisation ability across datasets. In this paper, we propose the Deep Facial Expression Vector ExtractoR (DeepFEVER), a new deep learning-based approach that learns a visual feature extractor general enough to be applied to any other facial emotion recognition task or dataset. DeepFEVER outperforms state-of-the-art results on the AffectNet and Google Facial Expression Comparison datasets. DeepFEVER’s extracted features also generalise extremely well to other datasets – even those unseen during training – namely, the Real-World Affective Faces (RAF) dataset.
Bae, Jin Hee, Kim, Minwoo, Lim, Joon S..  2021.  Emotion Detection and Analysis from Facial Image using Distance between Coordinates Feature. 2021 International Conference on Information and Communication Technology Convergence (ICTC). :494—497.
Facial expression recognition has long been established as a subject of continuous research in various fields. In this study, feature extraction was conducted by calculating the distance between facial landmarks in an image. The extracted features of the relationship between each landmark and analysis were used to classify five facial expressions. We increased the data and label reliability based on our labeling work with multiple observers. Additionally, faces were recognized from the original data, and landmark coordinates were extracted and used as features. A genetic algorithm was used to select features that were relatively more helpful for classification. We performed facial recognition classification and analysis using the method proposed in this study, which showed the validity and effectiveness of the proposed method.
Wang, Caixia, Wang, Zhihui, Cui, Dong.  2021.  Facial Expression Recognition with Attention Mechanism. 2021 14th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI). :1—6.
With the development of artificial intelligence, facial expression recognition (FER) has greatly improved performance in deep learning, but there is still a lot of room for improvement in the study of combining attention to focus the network on key parts of the face. For facial expression recognition, this paper designs a network model, which use spatial transformer network to transform the input image firstly, and then adding channel attention and spatial attention to the convolutional network. In addition, in this paper, the GELU activation function is used in the convolutional network, which improves the recognition rate of facial expressions to a certain extent.
Fallah, Zahra, Ebrahimpour-Komleh, Hossein, Mousavirad, Seyed Jalaleddin.  2021.  A Novel Hybrid Pyramid Texture-Based Facial Expression Recognition. 2021 5th International Conference on Pattern Recognition and Image Analysis (IPRIA). :1—6.
Automated analysis of facial expressions is one of the most interesting and challenging problems in many areas such as human-computer interaction. Facial images are affected by many factors, such as intensity, pose and facial expressions. These factors make facial expression recognition problem a challenge. The aim of this paper is to propose a new method based on the pyramid local binary pattern (PLBP) and the pyramid local phase quantization (PLPQ), which are the extension of the local binary pattern (LBP) and the local phase quantization (LPQ) as two methods for extracting texture features. LBP operator is used to extract LBP feature in the spatial domain and LPQ operator is used to extract LPQ feature in the frequency domain. The combination of features in spatial and frequency domains can provide important information in both domains. In this paper, PLBP and PLPQ operators are separately used to extract features. Then, these features are combined to create a new feature vector. The advantage of pyramid transform domain is that it can recognize facial expressions efficiently and with high accuracy even for very low-resolution facial images. The proposed method is verified on the CK+ facial expression database. The proposed method achieves the recognition rate of 99.85% on CK+ database.
Sun, Lanxin, Dai, JunBo, Shen, Xunbing.  2021.  Facial emotion recognition based on LDA and Facial Landmark Detection. 2021 2nd International Conference on Artificial Intelligence and Education (ICAIE). :64—67.
Emotion recognition in the field of human-computer interaction refers to that the computer has the corresponding perceptual ability to predict the emotional state of human beings in advance by observing human expressions, behaviors and emotions, so as to ensure that computers can communicate emotionally with humans. The main research work of this paper is to extract facial image features by using Linear Discriminant Analysis (LDA) and Facial Landmark Detection after grayscale processing and cropping, and then compare the accuracy after emotion recognition and classification to determine which feature extraction method is more effective. The test results show that the accuracy rate of emotion recognition in face images can reach 73.9% by using LDA method, and 84.5% by using Facial Landmark Detection method. Therefore, facial landmarks can be used to identify emotion in face images more accurately.
Cao, HongYuan, Qi, Chao.  2021.  Facial Expression Study Based on 3D Facial Emotion Recognition. 2021 20th International Conference on Ubiquitous Computing and Communications (IUCC/CIT/DSCI/SmartCNS). :375—381.
Teaching evaluation is an indispensable key link in the modern education model. Its purpose is to promote learners' cognitive and non-cognitive development, especially emotional development. However, today's education has increasingly neglected the emotional process of learners' learning. Therefore, a method of using machines to analyze the emotional changes of learners during learning has been proposed. At present, most of the existing emotion recognition algorithms use the extraction of two-dimensional facial features from images to perform emotion prediction. Through research, it is found that the recognition rate of 2D facial feature extraction is not optimal, so this paper proposes an effective the algorithm obtains a single two-dimensional image from the input end and constructs a three-dimensional face model from the output end, thereby using 3D facial information to estimate the continuous emotion of the dimensional space and applying this method to an online learning system. Experimental results show that the algorithm has strong robustness and recognition ability.
Siyaka, Hassan Opotu, Owolabi, Olumide, Bisallah, I. Hashim.  2021.  A New Facial Image Deviation Estimation and Image Selection Algorithm (Fide-Isa) for Facial Image Recognition Systems: The Mathematical Models. 2021 1st International Conference on Multidisciplinary Engineering and Applied Science (ICMEAS). :1—7.
Deep learning models have been successful and shown to perform better in terms of accuracy and efficiency for facial recognition applications. However, they require huge amount of data samples that were well annotated to be successful. Their data requirements have led to some complications which include increased processing demands of the systems where such systems were to be deployed. Reducing the training sample sizes of deep learning models is still an open problem. This paper proposes the reduction of the number of samples required by the convolutional neutral network used in training a facial recognition system using a new Facial Image Deviation Estimation and Image Selection Algorithm (FIDE-ISA). The algorithm was used to select appropriate facial image training samples incrementally based on their facial deviation. This will reduce the need for huge dataset in training deep learning models. Preliminary results indicated a 100% accuracy for models trained with 54 images (at least 3 images per individual) and above.
2022-06-30
Cao, Yu.  2021.  Digital Character CAPTCHA Recognition Using Convolution Network. 2021 2nd International Conference on Computing and Data Science (CDS). :130—135.
Completely Automated Public Turing test to tell Computers and Humans Apart (CAPTCHA) is a type of automatic program to determine whether the user is human or not. The most common type of CAPTCHA is a kind of message interpretation by twisting the letters and adding slight noises in the background, plays a role of verification code. In this paper, we will introduce the basis of Convolutional Neural Network first. Then based on the handwritten digit recognition using CNN, we will develop a network for CAPTCHA image recognition.
Jadhav, Mohit, Kulkarni, Nupur, Walhekar, Omkar.  2021.  Doodling Based CAPTCHA Authentication System. 2021 Asian Conference on Innovation in Technology (ASIANCON). :1—5.
CAPTCHA (Completely Automated Public Turing Test to tell Computers and Humans Apart) is a widely used challenge-measures to distinguish humans and computer automated programs apart. Several existing CAPTCHAs are reliable for normal users, whereas visually impaired users face a lot of problems with the CAPTCHA authentication process. CAPTCHAs such as Google reCAPTCHA alternatively provides audio CAPTCHA, but many users find it difficult to decipher due to noise, language barrier, and accent of the audio of the CAPTCHA. Existing CAPTCHA systems lack user satisfaction on smartphones thus limiting its use. Our proposed system potentially solves the problem faced by visually impaired users during the process of CAPTCHA authentication. Also, our system makes the authentication process generic across users as well as platforms.
Kumar, Ashwani, Singh, Aditya Pratap.  2021.  Contour Based Deep Learning Engine to Solve CAPTCHA. 2021 7th International Conference on Advanced Computing and Communication Systems (ICACCS). 1:723—727.
A 'Completely Automated Public Turing test to tell Computers and Humans Apart' or better known as CAPTCHA is a image based test used to determine the authenticity of a user (ie. whether the user is human or not). In today's world, almost all the web services, such as online shopping sites, require users to solve CAPTCHAs that must be read and typed correctly. The challenge is that recognizing the CAPTCHAs is a relatively easy task for humans, but it is still hard to solve for computers. Ideally, a well-designed CAPTCHA should be solvable by humans at least 90% of the time, while programs using appropriate resources should succeed in less than 0.01% of the cases. In this paper, a deep neural network architecture is presented to extract text from CAPTCHA images on various platforms. The central theme of the paper is to develop an efficient & intelligent model that converts image-based CAPTCHA to text. We used convolutional neural network based architecture design instead of the traditional methods of CAPTCHA detection using image processing segmentation modules. The model consists of seven layers to efficiently correlate image features to the output character sequence. We tried a wide variety of configurations, including various loss and activation functions. We generated our own images database and the efficacy of our model was proven by the accuracy levels of 99.7%.
2022-05-06
Wotawa, Franz, Klampfl, Lorenz, Jahaj, Ledio.  2021.  A framework for the automation of testing computer vision systems. 2021 IEEE/ACM International Conference on Automation of Software Test (AST). :121–124.
Vision systems, i.e., systems that enable the detection and tracking of objects in images, have gained substantial importance over the past decades. They are used in quality assurance applications, e.g., for finding surface defects in products during manufacturing, surveillance, but also automated driving, requiring reliable behavior. Interestingly, there is only little work on quality assurance and especially testing of vision systems in general. In this paper, we contribute to the area of testing vision software, and present a framework for the automated generation of tests for systems based on vision and image recognition with the focus on easy usage, uniform usability and expandability. The framework makes use of existing libraries for modifying the original images and to obtain similarities between the original and modified images. We show how such a framework can be used for testing a particular industrial application on identifying defects on riblet surfaces and present preliminary results from the image classification domain.
2022-03-09
Gong, Peiyong, Zheng, Kai, Jiang, Yi, Liu, Jia.  2021.  Water Surface Object Detection Based on Neural Style Learning Algorithm. 2021 40th Chinese Control Conference (CCC). :8539—8543.
In order to detect the objects on the water surface, a neural style learning algorithm is proposed in this paper. The algorithm uses the Gram matrix of a pre-trained convolutional neural network to represent the style of the texture in the image, which is originally used for image style transfer. The objects on the water surface can be easily distinguished by the difference in their styles of the image texture. The algorithm is tested on the dataset of the Airbus Ship Detection Challenge on Kaggle. Compared to the other water surface object detection algorithms, the proposed algorithm has a good precision of 0.925 with recall equals to 0.86.
2022-03-08
Tian, Qian, Song, Qishun, Wang, Hongbo, Hu, Zhihong, Zhu, Siyu.  2021.  Verification Code Recognition Based on Convolutional Neural Network. 2021 IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC). 4:1947—1950.

Verification code recognition system based on convolutional neural network. In order to strengthen the network security defense work, this paper proposes a novel verification code recognition system based on convolutional neural network. The system combines Internet technology and big data technology, combined with advanced captcha technology, can prevent hackers from brute force cracking behavior to a certain extent. In addition, the system combines convolutional neural network, which makes the verification code combine numbers and letters, which improves the complexity of the verification code and the security of the user account. Based on this, the system uses threshold segmentation method and projection positioning method to construct an 8-layer convolutional neural network model, which enhances the security of the verification code input link. The research results show that the system can enhance the complexity of captcha, improve the recognition rate of captcha, and improve the security of user accounting.

2022-03-01
Li, Dong, Jiao, Yiwen, Ge, Pengcheng, Sun, Kuanfei, Gao, Zefu, Mao, Feilong.  2021.  Classification Coding and Image Recognition Based on Pulse Neural Network. 2021 IEEE International Conference on Artificial Intelligence and Industrial Design (AIID). :260–265.
Based on the third generation neural network spiking neural network, this paper optimizes and improves a classification and coding method, and proposes an image recognition method. Firstly, the read image is converted into a spike sequence, and then the spike sequence is encoded in groups and sent to the neurons in the spike neural network. After learning and training for many times, the quantization standard code is obtained. In this process, the spike sequence transformation matrix and dynamic weight matrix are obtained, and the unclassified data are output through the same matrix for image recognition and classification. Simulation results show that the above methods can get correct coding and preliminary recognition classification, and the spiking neural network can be applied.
2022-01-25
Kozlova, Liudmila P., Kozlova, Olga A..  2021.  Expanding Space with Augmented Reality. 2021 IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (ElConRus). :965—967.
Replacing real life with the virtual space has long ceased to be a theory. Among the whole variety of visualization, systems that allow projecting non-existent objects into real-world space are especially distinguished. Thus, augmented reality technology has found its application in many different fields. The article discusses the general concepts and principles of building augmented reality systems.
2021-11-29
Li, Taojin, Lei, Songgui, Wang, Wei, Wang, Qingli.  2020.  Research on MR virtual scene location method based on image recognition. 2020 International Conference on Information Science, Parallel and Distributed Systems (ISPDS). :109–113.
In order to solve the problem of accurate positioning of mixed reality virtual scene in physical space, this paper, firstly, analyzes the positioning principle of mixed reality virtual scene. Secondly, based on the comparison among the three developer kits: ARToolKit, ARTag, and Vuforia and two image optimization algorithms: AHE and ACE, it makes sure to use Vuforia development tool to complete the signature-based tracking and registration task, and use ACE algorithm to optimize the signature-based image. It improves the efficiency, stability and accuracy of image recognition registration. Then the multi-target recognition and registration technology is used to realize the multi-location of virtual scene. Finally, Hololens glasses are used as the hardware carrier to verify the above method. The experimental results show that the above method not only realizes the precise location of MR virtual scene based on image recognition, but also ensures the absolute position of the virtual model in the real space, bringing users a more real virtual experience. Keywords-mixed reality, multi-person collaboration, virtual positioning, gesture interaction.
2021-08-31
Kim, Young-Sae, Han, Jin-Hee, Kim, Geonwoo.  2020.  Design of an efficient image protection method based on QR code. 2020 International Conference on Information and Communication Technology Convergence (ICTC). :1448—1450.
This paper presents the design and the verification of an efficient image protection method based on the QR code, which is a type of two-dimensional barcode widely used in various fields. For this purpose, we design a new image protection system consisting of a secure image generator and a secure image recognizer. One adds a new pre-processing block to the typical QR code generator and the other combines the existing QR code reader with a new post-processing block. The new architecture provides image de-identification. It is also flexible, allowing the use of text-based compression and encryption. We have implemented prototype applications for verifying the functions of the secure image generator and those of the secure image recognizer. As a result, it is shown that the proposed architecture can be used as a good solution for image privacy protection, especially in offline environments.
2021-07-08
Li, Sichun, Jin, Xin, Yao, Sibing, Yang, Shuyu.  2020.  Underwater Small Target Recognition Based on Convolutional Neural Network. Global Oceans 2020: Singapore – U.S. Gulf Coast. :1—7.
With the increasingly extensive use of diver and unmanned underwater vehicle in military, it has posed a serious threat to the security of the national coastal area. In order to prevent the underwater diver's impact on the safety of water area, it is of great significance to identify underwater small targets in time to make early warning for it. In this paper, convolutional neural network is applied to underwater small target recognition. The recognition targets are diver, whale and dolphin. Due to the time-frequency spectrum can reflect the essential features of underwater target, convolutional neural network can learn a variety of features of the acoustic signal through the image processed by the time-frequency spectrum, time-frequency image is input to convolutional neural network to recognize the underwater small targets. According to the study of learning rate and pooling mode, the network parameters and structure suitable for underwater small target recognition in this paper are selected. The results of data processing show that the method can identify underwater small targets accurately.