期刊文献+
共找到16,944篇文章
< 1 2 250 >
每页显示 20 50 100
Source localization in signed networks with effective distance
1
作者 马志伟 孙蕾 +2 位作者 丁智国 黄宜真 胡兆龙 《Chinese Physics B》 SCIE EI CAS CSCD 2024年第2期577-585,共9页
While progress has been made in information source localization,it has overlooked the prevalent friend and adversarial relationships in social networks.This paper addresses this gap by focusing on source localization ... While progress has been made in information source localization,it has overlooked the prevalent friend and adversarial relationships in social networks.This paper addresses this gap by focusing on source localization in signed network models.Leveraging the topological characteristics of signed networks and transforming the propagation probability into effective distance,we propose an optimization method for observer selection.Additionally,by using the reverse propagation algorithm we present a method for information source localization in signed networks.Extensive experimental results demonstrate that a higher proportion of positive edges within signed networks contributes to more favorable source localization,and the higher the ratio of propagation rates between positive and negative edges,the more accurate the source localization becomes.Interestingly,this aligns with our observation that,in reality,the number of friends tends to be greater than the number of adversaries,and the likelihood of information propagation among friends is often higher than among adversaries.In addition,the source located at the periphery of the network is not easy to identify.Furthermore,our proposed observer selection method based on effective distance achieves higher operational efficiency and exhibits higher accuracy in information source localization,compared with three strategies for observer selection based on the classical full-order neighbor coverage. 展开更多
关键词 complex networks signed networks source localization effective distance
原文传递
Recent Advances on Deep Learning for Sign Language Recognition
2
作者 Yanqiong Zhang Xianwei Jiang 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第6期2399-2450,共52页
Sign language,a visual-gestural language used by the deaf and hard-of-hearing community,plays a crucial role in facilitating communication and promoting inclusivity.Sign language recognition(SLR),the process of automa... Sign language,a visual-gestural language used by the deaf and hard-of-hearing community,plays a crucial role in facilitating communication and promoting inclusivity.Sign language recognition(SLR),the process of automatically recognizing and interpreting sign language gestures,has gained significant attention in recent years due to its potential to bridge the communication gap between the hearing impaired and the hearing world.The emergence and continuous development of deep learning techniques have provided inspiration and momentum for advancing SLR.This paper presents a comprehensive and up-to-date analysis of the advancements,challenges,and opportunities in deep learning-based sign language recognition,focusing on the past five years of research.We explore various aspects of SLR,including sign data acquisition technologies,sign language datasets,evaluation methods,and different types of neural networks.Convolutional Neural Networks(CNN)and Recurrent Neural Networks(RNN)have shown promising results in fingerspelling and isolated sign recognition.However,the continuous nature of sign language poses challenges,leading to the exploration of advanced neural network models such as the Transformer model for continuous sign language recognition(CSLR).Despite significant advancements,several challenges remain in the field of SLR.These challenges include expanding sign language datasets,achieving user independence in recognition systems,exploring different input modalities,effectively fusing features,modeling co-articulation,and improving semantic and syntactic understanding.Additionally,developing lightweight network architectures for mobile applications is crucial for practical implementation.By addressing these challenges,we can further advance the field of deep learning for sign language recognition and improve communication for the hearing-impaired community. 展开更多
关键词 sign language recognition deep learning artificial intelligence computer vision gesture recognition
下载PDF
Japanese Sign Language Recognition by Combining Joint Skeleton-Based Handcrafted and Pixel-Based Deep Learning Features with Machine Learning Classification
3
作者 Jungpil Shin Md.Al Mehedi Hasan +2 位作者 Abu Saleh Musa Miah Kota Suzuki Koki Hirooka 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第6期2605-2625,共21页
Sign language recognition is vital for enhancing communication accessibility among the Deaf and hard-of-hearing communities.In Japan,approximately 360,000 individualswith hearing and speech disabilities rely on Japane... Sign language recognition is vital for enhancing communication accessibility among the Deaf and hard-of-hearing communities.In Japan,approximately 360,000 individualswith hearing and speech disabilities rely on Japanese Sign Language(JSL)for communication.However,existing JSL recognition systems have faced significant performance limitations due to inherent complexities.In response to these challenges,we present a novel JSL recognition system that employs a strategic fusion approach,combining joint skeleton-based handcrafted features and pixel-based deep learning features.Our system incorporates two distinct streams:the first stream extracts crucial handcrafted features,emphasizing the capture of hand and body movements within JSL gestures.Simultaneously,a deep learning-based transfer learning stream captures hierarchical representations of JSL gestures in the second stream.Then,we concatenated the critical information of the first stream and the hierarchy of the second stream features to produce the multiple levels of the fusion features,aiming to create a comprehensive representation of the JSL gestures.After reducing the dimensionality of the feature,a feature selection approach and a kernel-based support vector machine(SVM)were used for the classification.To assess the effectiveness of our approach,we conducted extensive experiments on our Lab JSL dataset and a publicly available Arabic sign language(ArSL)dataset.Our results unequivocally demonstrate that our fusion approach significantly enhances JSL recognition accuracy and robustness compared to individual feature sets or traditional recognition methods. 展开更多
关键词 Japanese sign Language(JSL) hand gesture recognition geometric feature distance feature angle feature GoogleNet
下载PDF
A Survey on Chinese Sign Language Recognition:From Traditional Methods to Artificial Intelligence
4
作者 Xianwei Jiang Yanqiong Zhang +1 位作者 Juan Lei Yudong Zhang 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第7期1-40,共40页
Research on Chinese Sign Language(CSL)provides convenience and support for individuals with hearing impairments to communicate and integrate into society.This article reviews the relevant literature on Chinese Sign La... Research on Chinese Sign Language(CSL)provides convenience and support for individuals with hearing impairments to communicate and integrate into society.This article reviews the relevant literature on Chinese Sign Language Recognition(CSLR)in the past 20 years.Hidden Markov Models(HMM),Support Vector Machines(SVM),and Dynamic Time Warping(DTW)were found to be the most commonly employed technologies among traditional identificationmethods.Benefiting from the rapid development of computer vision and artificial intelligence technology,Convolutional Neural Networks(CNN),3D-CNN,YOLO,Capsule Network(CapsNet)and various deep neural networks have sprung up.Deep Neural Networks(DNNs)and their derived models are integral tomodern artificial intelligence recognitionmethods.In addition,technologies thatwerewidely used in the early days have also been integrated and applied to specific hybrid models and customized identification methods.Sign language data collection includes acquiring data from data gloves,data sensors(such as Kinect,LeapMotion,etc.),and high-definition photography.Meanwhile,facial expression recognition,complex background processing,and 3D sign language recognition have also attracted research interests among scholars.Due to the uniqueness and complexity of Chinese sign language,accuracy,robustness,real-time performance,and user independence are significant challenges for future sign language recognition research.Additionally,suitable datasets and evaluation criteria are also worth pursuing. 展开更多
关键词 Chinese sign Language Recognition deep neural networks artificial intelligence transfer learning hybrid network models
下载PDF
Constructing Representative Collective Signature Protocols Using The GOST R34.10-1994 Standard
5
作者 Tuan Nguyen Kim Duy Ho Ngoc Nikolay A.Moldovyan 《Computers, Materials & Continua》 SCIE EI 2023年第1期1475-1491,共17页
The representative collective digital signature,which was suggested by us,is built based on combining the advantages of group digital signature and collective digital signature.This collective digital signature schema... The representative collective digital signature,which was suggested by us,is built based on combining the advantages of group digital signature and collective digital signature.This collective digital signature schema helps to create a unique digital signature that deputizes a collective of people representing different groups of signers and may also include personal signers.The advantage of the proposed collective signature is that it can be built based on most of the well-known difficult problems such as the factor analysis,the discrete logarithm and finding modulo roots of large prime numbers and the current digital signature standards of the United States and Russian Federation.In this paper,we use the discrete logarithmic problem on prime finite fields,which has been implemented in the GOST R34.10-1994 digital signature standard,to build the proposed collective signature protocols.These protocols help to create collective signatures:Guaranteed internal integrity and fixed size,independent of the number of members involved in forming the signature.The signature built in this study,consisting of 3 components(U,R,S),stores the information of all relevant signers in the U components,thus tracking the signer and against the“disclaim of liability”of the signer later is possible.The idea of hiding the signer’s public key is also applied in the proposed protocols.This makes it easy for the signing group representative to specify which members are authorized to participate in the signature creation process. 展开更多
关键词 signing collective signing group discrete logarithm group signature collective signature GOST standards
下载PDF
Existence of Monotone Positive Solution for a Fourth-Order Three-Point BVP with Sign-Changing Green’s Function
6
作者 Junrui Yue Yun Zhang Qingyue Bai 《Open Journal of Applied Sciences》 2024年第1期63-69,共7页
This paper is concerned with the following fourth-order three-point boundary value problem , where , we discuss the existence of positive solutions to the above problem by applying to the fixed point theory in cones a... This paper is concerned with the following fourth-order three-point boundary value problem , where , we discuss the existence of positive solutions to the above problem by applying to the fixed point theory in cones and iterative technique. 展开更多
关键词 Fourth-Order Three-Point Boundary Value Problem sign-Changing Green’s Function Fixed Point Index Iterative Technique Monotone Positive Solution EXISTENCE
下载PDF
A Novel Action Transformer Network for Hybrid Multimodal Sign Language Recognition
7
作者 Sameena Javaid Safdar Rizvi 《Computers, Materials & Continua》 SCIE EI 2023年第1期523-537,共15页
Sign language fills the communication gap for people with hearing and speaking ailments.It includes both visual modalities,manual gestures consisting of movements of hands,and non-manual gestures incorporating body mo... Sign language fills the communication gap for people with hearing and speaking ailments.It includes both visual modalities,manual gestures consisting of movements of hands,and non-manual gestures incorporating body movements including head,facial expressions,eyes,shoulder shrugging,etc.Previously both gestures have been detected;identifying separately may have better accuracy,butmuch communicational information is lost.Aproper sign language mechanism is needed to detect manual and non-manual gestures to convey the appropriate detailed message to others.Our novel proposed system contributes as Sign LanguageAction Transformer Network(SLATN),localizing hand,body,and facial gestures in video sequences.Here we are expending a Transformer-style structural design as a“base network”to extract features from a spatiotemporal domain.Themodel impulsively learns to track individual persons and their action context inmultiple frames.Furthermore,a“head network”emphasizes hand movement and facial expression simultaneously,which is often crucial to understanding sign language,using its attention mechanism for creating tight bounding boxes around classified gestures.The model’s work is later compared with the traditional identification methods of activity recognition.It not only works faster but achieves better accuracy as well.Themodel achieves overall 82.66%testing accuracy with a very considerable performance of computation with 94.13 Giga-Floating Point Operations per Second(G-FLOPS).Another contribution is a newly created dataset of Pakistan Sign Language forManual and Non-Manual(PkSLMNM)gestures. 展开更多
关键词 sign language gesture recognition manual signs non-manual signs action transformer network
下载PDF
A Robust Model for Translating Arabic Sign Language into Spoken Arabic Using Deep Learning
8
作者 Khalid M.O.Nahar Ammar Almomani +1 位作者 Nahlah Shatnawi Mohammad Alauthman 《Intelligent Automation & Soft Computing》 SCIE 2023年第8期2037-2057,共21页
This study presents a novel and innovative approach to auto-matically translating Arabic Sign Language(ATSL)into spoken Arabic.The proposed solution utilizes a deep learning-based classification approach and the trans... This study presents a novel and innovative approach to auto-matically translating Arabic Sign Language(ATSL)into spoken Arabic.The proposed solution utilizes a deep learning-based classification approach and the transfer learning technique to retrain 12 image recognition models.The image-based translation method maps sign language gestures to corre-sponding letters or words using distance measures and classification as a machine learning technique.The results show that the proposed model is more accurate and faster than traditional image-based models in classifying Arabic-language signs,with a translation accuracy of 93.7%.This research makes a significant contribution to the field of ATSL.It offers a practical solution for improving communication for individuals with special needs,such as the deaf and mute community.This work demonstrates the potential of deep learning techniques in translating sign language into natural language and highlights the importance of ATSL in facilitating communication for individuals with disabilities. 展开更多
关键词 sign language deep learning transfer learning machine learning automatic translation of sign language natural language processing Arabic sign language
下载PDF
Traffic Sign Recognition for Autonomous Vehicle Using Optimized YOLOv7 and Convolutional Block Attention Module
9
作者 P.Kuppusamy M.Sanjay +1 位作者 P.V.Deepashree C.Iwendi 《Computers, Materials & Continua》 SCIE EI 2023年第10期445-466,共22页
The infrastructure and construction of roads are crucial for the economic and social development of a region,but traffic-related challenges like accidents and congestion persist.Artificial Intelligence(AI)and Machine ... The infrastructure and construction of roads are crucial for the economic and social development of a region,but traffic-related challenges like accidents and congestion persist.Artificial Intelligence(AI)and Machine Learning(ML)have been used in road infrastructure and construction,particularly with the Internet of Things(IoT)devices.Object detection in Computer Vision also plays a key role in improving road infrastructure and addressing trafficrelated problems.This study aims to use You Only Look Once version 7(YOLOv7),Convolutional Block Attention Module(CBAM),the most optimized object-detection algorithm,to detect and identify traffic signs,and analyze effective combinations of adaptive optimizers like Adaptive Moment estimation(Adam),Root Mean Squared Propagation(RMSprop)and Stochastic Gradient Descent(SGD)with the YOLOv7.Using a portion of German traffic signs for training,the study investigates the feasibility of adopting smaller datasets while maintaining high accuracy.The model proposed in this study not only improves traffic safety by detecting traffic signs but also has the potential to contribute to the rapid development of autonomous vehicle systems.The study results showed an impressive accuracy of 99.7%when using a batch size of 8 and the Adam optimizer.This high level of accuracy demonstrates the effectiveness of the proposed model for the image classification task of traffic sign recognition. 展开更多
关键词 Object detection traffic sign detection YOLOv7 convolutional block attention module road sign detection ADAM
下载PDF
Alphabet-Level Indian Sign Language Translation to Text Using Hybrid-AO Thresholding with CNN
10
作者 Seema Sabharwal Priti Singla 《Intelligent Automation & Soft Computing》 SCIE 2023年第9期2567-2582,共16页
Sign language is used as a communication medium in the field of trade,defence,and in deaf-mute communities worldwide.Over the last few decades,research in the domain of translation of sign language has grown and becom... Sign language is used as a communication medium in the field of trade,defence,and in deaf-mute communities worldwide.Over the last few decades,research in the domain of translation of sign language has grown and become more challenging.This necessitates the development of a Sign Language Translation System(SLTS)to provide effective communication in different research domains.In this paper,novel Hybrid Adaptive Gaussian Thresholding with Otsu Algorithm(Hybrid-AO)for image segmentation is proposed for the translation of alphabet-level Indian Sign Language(ISLTS)with a 5-layer Convolution Neural Network(CNN).The focus of this paper is to analyze various image segmentation(Canny Edge Detection,Simple Thresholding,and Hybrid-AO),pooling approaches(Max,Average,and Global Average Pooling),and activation functions(ReLU,Leaky ReLU,and ELU).5-layer CNN with Max pooling,Leaky ReLU activation function,and Hybrid-AO(5MXLR-HAO)have outperformed other frameworks.An open-access dataset of ISL alphabets with approx.31 K images of 26 classes have been used to train and test the model.The proposed framework has been developed for translating alphabet-level Indian Sign Language into text.The proposed framework attains 98.95%training accuracy,98.05%validation accuracy,and 0.0721 training loss and 0.1021 validation loss and the perfor-mance of the proposed system outperforms other existing systems. 展开更多
关键词 sign language translation CNN THRESHOLDING Indian sign language
下载PDF
Rotation,Translation and Scale Invariant Sign Word Recognition Using Deep Learning 被引量:2
11
作者 Abu Saleh Musa Miah Jungpil Shin +2 位作者 Md.Al Mehedi Hasan Md Abdur Rahim Yuichi Okuyama 《Computer Systems Science & Engineering》 SCIE EI 2023年第3期2521-2536,共16页
Communication between people with disabilities and people who do not understand sign language is a growing social need and can be a tedious task.One of the main functions of sign language is to communicate with each o... Communication between people with disabilities and people who do not understand sign language is a growing social need and can be a tedious task.One of the main functions of sign language is to communicate with each other through hand gestures.Recognition of hand gestures has become an important challenge for the recognition of sign language.There are many existing models that can produce a good accuracy,but if the model test with rotated or translated images,they may face some difficulties to make good performance accuracy.To resolve these challenges of hand gesture recognition,we proposed a Rotation,Translation and Scale-invariant sign word recognition system using a convolu-tional neural network(CNN).We have followed three steps in our work:rotated,translated and scaled(RTS)version dataset generation,gesture segmentation,and sign word classification.Firstly,we have enlarged a benchmark dataset of 20 sign words by making different amounts of Rotation,Translation and Scale of the ori-ginal images to create the RTS version dataset.Then we have applied the gesture segmentation technique.The segmentation consists of three levels,i)Otsu Thresholding with YCbCr,ii)Morphological analysis:dilation through opening morphology and iii)Watershed algorithm.Finally,our designed CNN model has been trained to classify the hand gesture as well as the sign word.Our model has been evaluated using the twenty sign word dataset,five sign word dataset and the RTS version of these datasets.We achieved 99.30%accuracy from the twenty sign word dataset evaluation,99.10%accuracy from the RTS version of the twenty sign word evolution,100%accuracy from thefive sign word dataset evaluation,and 98.00%accuracy from the RTS versionfive sign word dataset evolution.Furthermore,the influence of our model exists in competitive results with state-of-the-art methods in sign word recognition. 展开更多
关键词 sign word recognition convolution neural network(cnn) rotation translation and scaling(rts) otsu segmentation
下载PDF
Simulation based on a modified social force model for sensitivity to emergency signs in subway station 被引量:1
12
作者 蔡征宇 周汝 +2 位作者 崔银锴 王妍 蒋军成 《Chinese Physics B》 SCIE EI CAS CSCD 2023年第2期175-183,共9页
The subway is the primary travel tool for urban residents in China. Due to the complex structure of the subway and high personnel density in rush hours, subway evacuation capacity is critical. The subway evacuation mo... The subway is the primary travel tool for urban residents in China. Due to the complex structure of the subway and high personnel density in rush hours, subway evacuation capacity is critical. The subway evacuation model is explored in this work by combining the improved social force model with the view radius using the Vicsek model. The pedestrians are divided into two categories based on different force models. The first category is sensitive pedestrians who have normal responses to emergency signs. The second category is insensitive pedestrians. By simulating different proportions of the insensitive pedestrians, we find that the escape time is directly proportional to the number of insensitive pedestrians and inversely proportional to the view radius. However, when the view radius is large enough, the escape time does not change significantly, and the evacuation of people in a small view radius environment tends to be integrated. With the improvement of view radius conditions, the escape time changes more obviously with the proportion of insensitive pedestrians. A new emergency sign layout is proposed, and the simulations show that the proposed layout can effectively reduce the escape time in a small view radius environment. However, the evacuation effect of the new escape sign layout on the large view radius environment is not apparent. In this case, the exit setting emerges as an additional factor affecting the escape time. 展开更多
关键词 modified social force model emergency evacuation insensitive pedestrians emergency signs layout
原文传递
A Light-Weight Deep Learning-Based Architecture for Sign Language Classification 被引量:1
13
作者 M.Daniel Nareshkumar B.Jaison 《Intelligent Automation & Soft Computing》 SCIE 2023年第3期3501-3515,共15页
With advancements in computing powers and the overall quality of images captured on everyday cameras,a much wider range of possibilities has opened in various scenarios.This fact has several implications for deaf and ... With advancements in computing powers and the overall quality of images captured on everyday cameras,a much wider range of possibilities has opened in various scenarios.This fact has several implications for deaf and dumb people as they have a chance to communicate with a greater number of people much easier.More than ever before,there is a plethora of info about sign language usage in the real world.Sign languages,and by extension the datasets available,are of two forms,isolated sign language and continuous sign language.The main difference between the two types is that in isolated sign language,the hand signs cover individual letters of the alphabet.In continuous sign language,entire words’hand signs are used.This paper will explore a novel deep learning architecture that will use recently published large pre-trained image models to quickly and accurately recognize the alphabets in the American Sign Language(ASL).The study will focus on isolated sign language to demonstrate that it is possible to achieve a high level of classification accuracy on the data,thereby showing that interpreters can be implemented in the real world.The newly proposed Mobile-NetV2 architecture serves as the backbone of this study.It is designed to run on end devices like mobile phones and infer signals(what does it infer)from images in a relatively short amount of time.With the proposed architecture in this paper,the classification accuracy of 98.77%in the Indian Sign Language(ISL)and American Sign Language(ASL)is achieved,outperforming the existing state-of-the-art systems. 展开更多
关键词 Deep learning machine learning CLASSIFICATION filters american sign language
下载PDF
Continuous Sign Language Recognition Based on Spatial-Temporal Graph Attention Network 被引量:1
14
作者 Qi Guo Shujun Zhang Hui Li 《Computer Modeling in Engineering & Sciences》 SCIE EI 2023年第3期1653-1670,共18页
Continuous sign language recognition(CSLR)is challenging due to the complexity of video background,hand gesture variability,and temporal modeling difficulties.This work proposes a CSLR method based on a spatialtempora... Continuous sign language recognition(CSLR)is challenging due to the complexity of video background,hand gesture variability,and temporal modeling difficulties.This work proposes a CSLR method based on a spatialtemporal graph attention network to focus on essential features of video series.The method considers local details of sign language movements by taking the information on joints and bones as inputs and constructing a spatialtemporal graph to reflect inter-frame relevance and physical connections between nodes.The graph-based multihead attention mechanism is utilized with adjacent matrix calculation for better local-feature exploration,and short-term motion correlation modeling is completed via a temporal convolutional network.We adopted BLSTM to learn the long-termdependence and connectionist temporal classification to align the word-level sequences.The proposed method achieves competitive results regarding word error rates(1.59%)on the Chinese Sign Language dataset and the mean Jaccard Index(65.78%)on the ChaLearn LAP Continuous Gesture Dataset. 展开更多
关键词 Continuous sign language recognition graph attention network bidirectional long short-term memory connectionist temporal classification
下载PDF
Traffic Sign Detection with Low Complexity for Intelligent Vehicles Based on Hybrid Features
15
作者 Sara Khalid Jamal Hussain Shah +2 位作者 Muhammad Sharif Muhammad Rafiq Gyu Sang Choi 《Computers, Materials & Continua》 SCIE EI 2023年第7期861-879,共19页
Globally traffic signs are used by all countries for healthier traffic flow and to protect drivers and pedestrians.Consequently,traffic signs have been of great importance for every civilized country,which makes resea... Globally traffic signs are used by all countries for healthier traffic flow and to protect drivers and pedestrians.Consequently,traffic signs have been of great importance for every civilized country,which makes researchers give more focus on the automatic detection of traffic signs.Detecting these traffic signs is challenging due to being in the dark,far away,partially occluded,and affected by the lighting or the presence of similar objects.An innovative traffic sign detection method for red and blue signs in color images is proposed to resolve these issues.This technique aimed to devise an efficient,robust and accurate approach.To attain this,initially,the approach presented a new formula,inspired by existing work,to enhance the image using red and green channels instead of blue,which segmented using a threshold calculated from the correlational property of the image.Next,a new set of features is proposed,motivated by existing features.Texture and color features are fused after getting extracted on the channel of Red,Green,and Blue(RGB),Hue,Saturation,and Value(HSV),and YCbCr color models of images.Later,the set of features is employed on different classification frameworks,from which quadratic support vector machine(SVM)outnumbered the others with an accuracy of 98.5%.The proposed method is tested on German Traffic Sign Detection Benchmark(GTSDB)images.The results are satisfactory when compared to the preceding work. 展开更多
关键词 Traffic sign detection intelligent systems COMPLEXITY VEHICLES color moments texture features
下载PDF
Arabic Sign Language Gesture Classification Using Deer Hunting Optimization with Machine Learning Model
16
作者 Badriyya B.Al-onazi Mohamed K.Nour +6 位作者 Hussain Alshahran Mohamed Ahmed Elfaki Mrim M.Alnfiai Radwa Marzouk Mahmoud Othman Mahir M.Sharif Abdelwahed Motwakel 《Computers, Materials & Continua》 SCIE EI 2023年第5期3413-3429,共17页
Sign language includes the motion of the arms and hands to communicate with people with hearing disabilities.Several models have been available in the literature for sign language detection and classification for enha... Sign language includes the motion of the arms and hands to communicate with people with hearing disabilities.Several models have been available in the literature for sign language detection and classification for enhanced outcomes.But the latest advancements in computer vision enable us to perform signs/gesture recognition using deep neural networks.This paper introduces an Arabic Sign Language Gesture Classification using Deer Hunting Optimization with Machine Learning(ASLGC-DHOML)model.The presented ASLGC-DHOML technique mainly concentrates on recognising and classifying sign language gestures.The presented ASLGC-DHOML model primarily pre-processes the input gesture images and generates feature vectors using the densely connected network(DenseNet169)model.For gesture recognition and classification,a multilayer perceptron(MLP)classifier is exploited to recognize and classify the existence of sign language gestures.Lastly,the DHO algorithm is utilized for parameter optimization of the MLP model.The experimental results of the ASLGC-DHOML model are tested and the outcomes are inspected under distinct aspects.The comparison analysis highlighted that the ASLGC-DHOML method has resulted in enhanced gesture classification results than other techniques with maximum accuracy of 92.88%. 展开更多
关键词 Machine learning sign language recognition multilayer perceptron deer hunting optimization densenet
下载PDF
Pre-Locator Incorporating Swin-Transformer Refined Classifier for Traffic Sign Recognition
17
作者 Qiang Luo Wenbin Zheng 《Intelligent Automation & Soft Computing》 SCIE 2023年第8期2227-2246,共20页
In the field of traffic sign recognition,traffic signs usually occupy very small areas in the input image.Most object detection algorithms directly reduce the original image to a specific size for the input model duri... In the field of traffic sign recognition,traffic signs usually occupy very small areas in the input image.Most object detection algorithms directly reduce the original image to a specific size for the input model during the detection process,which leads to the loss of small object information.Addi-tionally,classification tasks are more sensitive to information loss than local-ization tasks.This paper proposes a novel traffic sign recognition approach,in which a lightweight pre-locator network and a refined classification network are incorporated.The pre-locator network locates the sub-regions of the traffic signs from the original image,and the refined classification network performs the refinement recognition task in the sub-regions.Moreover,an innovative module(named SPP-ST)is proposed,which combines the Spatial Pyramid Pool module(SPP)and the Swin-Transformer module as a new feature extractor to learn the special spatial information of traffic sign effec-tively.Experimental results show that the proposed method is superior to the state-of-the-art methods(82.1 mAP achieved on 218 categories in the TT100k dataset,an improvement of 19.7 percentage points compared to the previous method).Moreover,both the result analysis and the output visualizations further demonstrate the effectiveness of our proposed method.The source code and datasets of this work are available at https://github.com/DijiesitelaQ/TSOD. 展开更多
关键词 Traffic sign RECOGNITION swin-transformer YOLOX small object
下载PDF
C2Net-YOLOv5: A Bidirectional Res2Net-Based Traffic Sign Detection Algorithm
18
作者 Xiujuan Wang Yiqi Tian +1 位作者 Kangfeng Zheng Chutong Liu 《Computers, Materials & Continua》 SCIE EI 2023年第11期1949-1965,共17页
Rapid advancement of intelligent transportation systems(ITS)and autonomous driving(AD)have shown the importance of accurate and efficient detection of traffic signs.However,certain drawbacks,such as balancing accuracy... Rapid advancement of intelligent transportation systems(ITS)and autonomous driving(AD)have shown the importance of accurate and efficient detection of traffic signs.However,certain drawbacks,such as balancing accuracy and real-time performance,hinder the deployment of traffic sign detection algorithms in ITS and AD domains.In this study,a novel traffic sign detection algorithm was proposed based on the bidirectional Res2Net architecture to achieve an improved balance between accuracy and speed.An enhanced backbone network module,called C2Net,which uses an upgraded bidirectional Res2Net,was introduced to mitigate information loss in the feature extraction process and to achieve information complementarity.Furthermore,a squeeze-and-excitation attention mechanism was incorporated within the channel attention of the architecture to perform channel-level feature correction on the input feature map,which effectively retains valuable features while removing non-essential features.A series of ablation experiments were conducted to validate the efficacy of the proposed methodology.The performance was evaluated using two distinct datasets:the Tsinghua-Tencent 100K and the CSUST Chinese traffic sign detection benchmark 2021.On the TT100K dataset,the method achieves precision,recall,and Map0.5 scores of 83.3%,79.3%,and 84.2%,respectively.Similarly,on the CCTSDB 2021 dataset,the method achieves precision,recall,and Map0.5 scores of 91.49%,73.79%,and 81.03%,respectively.Experimental results revealed that the proposed method had superior performance compared to conventional models,which includes the faster region-based convolutional neural network,single shot multibox detector,and you only look once version 5. 展开更多
关键词 Target detection traffic sign detection autonomous driving YOLOv5
下载PDF
Identification and Acknowledgment of Programmed Traffic Sign Utilizing Profound Convolutional Neural Organization
19
作者 P.Vigneshwaran N.Prasath +1 位作者 M.Islabudeen A.Arunand A.K.Sampath 《Intelligent Automation & Soft Computing》 SCIE 2023年第2期1527-1543,共17页
Traffic signs are basic security workplaces making the rounds,which expects a huge part in coordinating busy time gridlock direct,ensuring the pros-perity of the road and dealing with the smooth segment of vehicles and... Traffic signs are basic security workplaces making the rounds,which expects a huge part in coordinating busy time gridlock direct,ensuring the pros-perity of the road and dealing with the smooth segment of vehicles and indivi-duals by walking,etc.As a segment of the clever transportation structure,the acknowledgment of traffic signs is basic for the driving assistance system,traffic sign upkeep,self-administering driving,and various spaces.There are different assessments turns out achieved for traffic sign acknowledgment in the world.However,most of the works are only for explicit arrangements of traffic signs,for example,beyond what many would consider a possible sign.Traffic sign recognizable proof is generally seen as trying on account of various complexities,for example,extended establishments of traffic sign pictures.Two critical issues exist during the time spent identification(ID)and affirmation of traffic signals.Road signs are occasionally blocked not entirely by various vehicles and various articles are accessible in busy time gridlock scenes which make the signed acknowledgment hard and walkers,various vehicles,constructions,and loads up may frustrate the ID structure by plans like that of road signs.Also concealing information from traffic scene pictures is affected by moving light achieved by environment conditions,time(day-night),and shadowing.Traffic sign revelation and affirmation structure has two guideline sorts out:The essential stage incorpo-rates the traffic sign limitation and the resulting stage portrays the perceived traffic signs into a particular class. 展开更多
关键词 Traffic sign classifier convolution neural network image vehicle
下载PDF
A Deep Learning Model of Traffic Signs in Panoramic Images Detection
20
作者 Kha Tu Huynh Thi Phuong Linh Le +1 位作者 Muhammad Arif Thien Khai Tran 《Intelligent Automation & Soft Computing》 SCIE 2023年第7期401-418,共18页
To pursue the ideal of a safe high-tech society in a time when traffic accidents are frequent,the traffic signs detection system has become one of the necessary topics in recent years and in the future.The ultimate go... To pursue the ideal of a safe high-tech society in a time when traffic accidents are frequent,the traffic signs detection system has become one of the necessary topics in recent years and in the future.The ultimate goal of this research is to identify and classify the types of traffic signs in a panoramic image.To accomplish this goal,the paper proposes a new model for traffic sign detection based on the Convolutional Neural Network for com-prehensive traffic sign classification and Mask Region-based Convolutional Neural Networks(R-CNN)implementation for identifying and extracting signs in panoramic images.Data augmentation and normalization of the images are also applied to assist in classifying better even if old traffic signs are degraded,and considerably minimize the rates of discovering the extra boxes.The proposed model is tested on both the testing dataset and the actual images and gets 94.5%of the correct signs recognition rate,the classification rate of those signs discovered was 99.41%and the rate of false signs was only around 0.11. 展开更多
关键词 Deep learning convolutional neural network Mask R-CNN traffic signs detection
下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部