期刊文献+
共找到25,032篇文章
< 1 2 250 >
每页显示 20 50 100
Structured Multi-Head Attention Stock Index Prediction Method Based Adaptive Public Opinion Sentiment Vector
1
作者 Cheng Zhao Zhe Peng +2 位作者 Xuefeng Lan Yuefeng Cen Zuxin Wang 《Computers, Materials & Continua》 SCIE EI 2024年第1期1503-1523,共21页
The present study examines the impact of short-term public opinion sentiment on the secondary market,with a focus on the potential for such sentiment to cause dramatic stock price fluctuations and increase investment ... The present study examines the impact of short-term public opinion sentiment on the secondary market,with a focus on the potential for such sentiment to cause dramatic stock price fluctuations and increase investment risk.The quantification of investment sentiment indicators and the persistent analysis of their impact has been a complex and significant area of research.In this paper,a structured multi-head attention stock index prediction method based adaptive public opinion sentiment vector is proposed.The proposedmethod utilizes an innovative approach to transform numerous investor comments on social platforms over time into public opinion sentiment vectors expressing complex sentiments.It then analyzes the continuous impact of these vectors on the market through the use of aggregating techniques and public opinion data via a structured multi-head attention mechanism.The experimental results demonstrate that the public opinion sentiment vector can provide more comprehensive feedback on market sentiment than traditional sentiment polarity analysis.Furthermore,the multi-head attention mechanism is shown to improve prediction accuracy through attention convergence on each type of input information separately.Themean absolute percentage error(MAPE)of the proposedmethod is 0.463%,a reduction of 0.294% compared to the benchmark attention algorithm.Additionally,the market backtesting results indicate that the return was 24.560%,an improvement of 8.202% compared to the benchmark algorithm.These results suggest that themarket trading strategy based on thismethod has the potential to improve trading profits. 展开更多
关键词 Public opinion sentiment structured multi-head attention stock index prediction deep learning
下载PDF
Multi-Head Attention Spatial-Temporal Graph Neural Networks for Traffic Forecasting
2
作者 Xiuwei Hu Enlong Yu Xiaoyu Zhao 《Journal of Computer and Communications》 2024年第3期52-67,共16页
Accurate traffic prediction is crucial for an intelligent traffic system (ITS). However, the excessive non-linearity and complexity of the spatial-temporal correlation in traffic flow severely limit the prediction acc... Accurate traffic prediction is crucial for an intelligent traffic system (ITS). However, the excessive non-linearity and complexity of the spatial-temporal correlation in traffic flow severely limit the prediction accuracy of most existing models, which simply stack temporal and spatial modules and fail to capture spatial-temporal features effectively. To improve the prediction accuracy, a multi-head attention spatial-temporal graph neural network (MSTNet) is proposed in this paper. First, the traffic data is decomposed into unique time spans that conform to positive rules, and valuable traffic node attributes are mined through an adaptive graph structure. Second, time and spatial features are captured using a multi-head attention spatial-temporal module. Finally, a multi-step prediction module is used to achieve future traffic condition prediction. Numerical experiments were conducted on an open-source dataset, and the results demonstrate that MSTNet performs well in spatial-temporal feature extraction and achieves more positive forecasting results than the baseline methods. 展开更多
关键词 Traffic Prediction Intelligent Traffic System multi-head attention Graph Neural Networks
下载PDF
Discharge Summaries Based Sentiment Detection Using Multi-Head Attention and CNN-BiGRU
3
作者 Samer Abdulateef Waheeb 《Computer Systems Science & Engineering》 SCIE EI 2023年第7期981-998,共18页
Automatic extraction of the patient’s health information from the unstructured data concerning the discharge summary remains challenging.Discharge summary related documents contain various aspects of the patient heal... Automatic extraction of the patient’s health information from the unstructured data concerning the discharge summary remains challenging.Discharge summary related documents contain various aspects of the patient health condition to examine the quality of treatment and thereby help improve decision-making in the medical field.Using a sentiment dictionary and feature engineering,the researchers primarily mine semantic text features.However,choosing and designing features requires a lot of manpower.The proposed approach is an unsupervised deep learning model that learns a set of clusters embedded in the latent space.A composite model including Active Learning(AL),Convolutional Neural Network(CNN),BiGRU,and Multi-Attention,called ACBMA in this research,is designed to measure the quality of treatment based on discharge summaries text sentiment detection.CNN is utilized for extracting the set of local features of text vectors.Then BiGRU network was utilized to extract the text’s global features to solve the issues that a single CNN cannot obtain global semantic information and the traditional Recurrent Neural Network(RNN)gradient disappearance.Experiments prove that the ACBMA method can demonstrate the effectiveness of the suggested method,achieve comparable results to state-of-arts methods in sentiment detection,and outperform them with accurate benchmarks.Finally,several algorithm studies ultimately determined that the ACBMA method is more precise for discharge summaries sentiment analysis. 展开更多
关键词 Sentiment analysis LEXICON discharge summaries active learning multi-head attention mechanism
下载PDF
Using Recurrent Neural Network Structure and Multi-Head Attention with Convolution for Fraudulent Phone Text Recognition
4
作者 Junjie Zhou Hongkui Xu +3 位作者 Zifeng Zhang Jiangkun Lu Wentao Guo Zhenye Li 《Computer Systems Science & Engineering》 SCIE EI 2023年第8期2277-2297,共21页
Fraud cases have been a risk in society and people’s property security has been greatly threatened.In recent studies,many promising algorithms have been developed for social media offensive text recognition as well a... Fraud cases have been a risk in society and people’s property security has been greatly threatened.In recent studies,many promising algorithms have been developed for social media offensive text recognition as well as sentiment analysis.These algorithms are also suitable for fraudulent phone text recognition.Compared to these tasks,the semantics of fraudulent words are more complex and more difficult to distinguish.Recurrent Neural Networks(RNN),the variants ofRNN,ConvolutionalNeuralNetworks(CNN),and hybrid neural networks to extract text features are used by most text classification research.However,a single network or a simple network combination cannot obtain rich characteristic knowledge of fraudulent phone texts relatively.Therefore,a new model is proposed in this paper.In the fraudulent phone text,the knowledge that can be learned by the model includes the sequence structure of sentences,the correlation between words,the correlation of contextual semantics,the feature of keywords in sentences,etc.The new model combines a bidirectional Long-Short Term Memory Neural Network(BiLSTM)or a bidirectional Gate Recurrent United(BiGRU)and a Multi-Head attention mechanism module with convolution.A normalization layer is added after the output of the final hidden layer.BiLSTM or BiGRU is used to build the encoding and decoding layer.Multi-head attention mechanism module with convolution(MHAC)enhances the ability of the model to learn global interaction information and multi-granularity local interaction information in fraudulent sentences.A fraudulent phone text dataset is produced by us in this paper.The THUCNews data sets and fraudulent phone text data sets are used in experiments.Experiment results show that compared with the baseline model,the proposed model(LMHACL)has the best experiment results in terms of Accuracy,Precision,Recall,and F1 score on the two data sets.And the performance indexes on fraudulent phone text data sets are all above 0.94. 展开更多
关键词 BiLSTM BiGRU multi-head attention mechanism CNN
下载PDF
基于Multi-head Attention和Bi-LSTM的实体关系分类 被引量:11
5
作者 刘峰 高赛 +1 位作者 于碧辉 郭放达 《计算机系统应用》 2019年第6期118-124,共7页
关系分类是自然语言处理领域的一项重要任务,能够为知识图谱的构建、问答系统和信息检索等提供技术支持.与传统关系分类方法相比较,基于神经网络和注意力机制的关系分类模型在各种关系分类任务中都获得了更出色的表现.以往的模型大多采... 关系分类是自然语言处理领域的一项重要任务,能够为知识图谱的构建、问答系统和信息检索等提供技术支持.与传统关系分类方法相比较,基于神经网络和注意力机制的关系分类模型在各种关系分类任务中都获得了更出色的表现.以往的模型大多采用单层注意力机制,特征表达相对单一.因此本文在已有研究基础上,引入多头注意力机制(Multi-head attention),旨在让模型从不同表示空间上获取关于句子更多层面的信息,提高模型的特征表达能力.同时在现有的词向量和位置向量作为网络输入的基础上,进一步引入依存句法特征和相对核心谓词依赖特征,其中依存句法特征包括当前词的依存关系值和所依赖的父节点位置,从而使模型进一步获取更多的文本句法信息.在SemEval-2010 任务8 数据集上的实验结果证明,该方法相较之前的深度学习模型,性能有进一步提高. 展开更多
关键词 关系分类 Bi-LSTM 句法特征 self-attention multi-head attention
下载PDF
Multi-Head Attention Graph Network for Few Shot Learning
6
作者 Baiyan Zhang Hefei Ling +5 位作者 Ping Li Qian Wang Yuxuan Shi Lei Wu Runsheng Wang Jialie Shen 《Computers, Materials & Continua》 SCIE EI 2021年第8期1505-1517,共13页
The majority of existing graph-network-based few-shot models focus on a node-similarity update mode.The lack of adequate information intensies the risk of overtraining.In this paper,we propose a novel Multihead Attent... The majority of existing graph-network-based few-shot models focus on a node-similarity update mode.The lack of adequate information intensies the risk of overtraining.In this paper,we propose a novel Multihead Attention Graph Network to excavate discriminative relation and fulll effective information propagation.For edge update,the node-level attention is used to evaluate the similarities between the two nodes and the distributionlevel attention extracts more in-deep global relation.The cooperation between those two parts provides a discriminative and comprehensive expression for edge feature.For node update,we embrace the label-level attention to soften the noise of irrelevant nodes and optimize the update direction.Our proposed model is veried through extensive experiments on two few-shot benchmark MiniImageNet and CIFAR-FS dataset.The results suggest that our method has a strong capability of noise immunity and quick convergence.The classication accuracy outperforms most state-of-the-art approaches. 展开更多
关键词 Few shot learning attention graph network
下载PDF
引入上下文信息和Attention Gate的GUS-YOLO遥感目标检测算法
7
作者 张华卫 张文飞 +2 位作者 蒋占军 廉敬 吴佰靖 《计算机科学与探索》 CSCD 2024年第2期453-464,共12页
目前基于通用YOLO系列的遥感目标检测算法存在并未充分利用图像的全局上下文信息,在特征融合金字塔部分并未充分考虑缩小融合特征之间的语义鸿沟、抑制冗余信息干扰的缺点。在结合YOLO算法优点的基础上提出GUS-YOLO算法,其拥有一个能够... 目前基于通用YOLO系列的遥感目标检测算法存在并未充分利用图像的全局上下文信息,在特征融合金字塔部分并未充分考虑缩小融合特征之间的语义鸿沟、抑制冗余信息干扰的缺点。在结合YOLO算法优点的基础上提出GUS-YOLO算法,其拥有一个能够充分利用全局上下文信息的骨干网络Global Backbone。除此之外,该算法在融合特征金字塔自顶向下的结构中引入Attention Gate模块,可以突出必要的特征信息,抑制冗余信息。另外,为Attention Gate模块设计了最佳的网络结构,提出了网络的特征融合结构U-Net。最后,为克服ReLU函数可能导致模型梯度不再更新的问题,该算法将Attention Gate模块的激活函数升级为可学习的SMU激活函数,提高模型鲁棒性。在NWPU VHR-10遥感数据集上,该算法相较于YOLOV7算法取得宽松指标mAP^(0.50)1.64个百分点和严格指标mAP^(0.75)9.39个百分点的性能提升。相较于目前主流的七种检测算法,该算法取得较好的检测性能。 展开更多
关键词 遥感图像 Global Backbone attention Gate SMU U-neck
下载PDF
Short-term load forecasting model based on gated recurrent unit and multi-head attention 被引量:2
8
作者 Li Hao Zhang Linghua +1 位作者 Tong Cheng Zhou Chenyang 《The Journal of China Universities of Posts and Telecommunications》 EI CSCD 2023年第3期25-31,共7页
Short-term load forecasting(STLF)plays a crucial role in the smart grid.However,it is challenging to capture the long-time dependence and the nonlinear relationship due to the comprehensive fluctuations of the electri... Short-term load forecasting(STLF)plays a crucial role in the smart grid.However,it is challenging to capture the long-time dependence and the nonlinear relationship due to the comprehensive fluctuations of the electrical load.In this paper,an STLF model based on gated recurrent unit and multi-head attention(GRU-MA)is proposed to address the aforementioned problems.The proposed model accommodates the time series and nonlinear relationship of load data through gated recurrent unit(GRU)and exploits multi-head attention(MA)to learn the decisive features and long-term dependencies.Additionally,the proposed model is compared with the support vector regression(SVR)model,the recurrent neural network and multi-head attention(RNN-MA)model,the long short-term memory and multi-head attention(LSTM-MA)model,the GRU model,and the temporal convolutional network(TCN)model using the public dataset of the Global Energy Forecasting Competition 2014(GEFCOM2014).The results demonstrate that the GRU-MA model has the best prediction accuracy. 展开更多
关键词 deep learning short-term load forecasting(STLF) gated recurrent unit(GRU) multi-head attention(MA)
原文传递
Fiber communication receiver models based on the multi-head attention mechanism
9
作者 臧裕斌 于振明 +3 位作者 徐坤 陈明华 杨四刚 陈宏伟 《Chinese Optics Letters》 SCIE EI CAS CSCD 2023年第3期29-34,共6页
In this paper,an artificial-intelligence-based fiber communication receiver model is put forward.With the multi-head attention mechanism it contains,this model can extract crucial patterns and map the transmitted sign... In this paper,an artificial-intelligence-based fiber communication receiver model is put forward.With the multi-head attention mechanism it contains,this model can extract crucial patterns and map the transmitted signals into the bit stream.Once appropriately trained,it can obtain the ability to restore the information from the signals whose transmission distances range from 0 to 100 km,signal-to-noise ratios range from 0 to 20 dB,modulation formats range from OOK to PAM4,and symbol rates range from 10 to 40 GBaud.The validity of the model is numerically demonstrated via MATLAB and Pytorch scenarios and compared with traditional communication receivers. 展开更多
关键词 fiber receiver model neural networks multi-head attention mechanism
原文传递
融合RoBERTa-GCN-Attention的隐喻识别与情感分类模型
10
作者 杨春霞 韩煜 +1 位作者 桂强 陈启岗 《小型微型计算机系统》 CSCD 2024年第3期576-583,共8页
在隐喻识别与隐喻情感分类任务的联合研究中,现有多任务学习模型存在对隐喻语料中的上下文语义信息和句法结构信息提取不够准确,并且缺乏对粗细两种粒度信息同时捕捉的问题.针对第1个问题,首先改进了传统的RoBERTa模型,在原有的自注意... 在隐喻识别与隐喻情感分类任务的联合研究中,现有多任务学习模型存在对隐喻语料中的上下文语义信息和句法结构信息提取不够准确,并且缺乏对粗细两种粒度信息同时捕捉的问题.针对第1个问题,首先改进了传统的RoBERTa模型,在原有的自注意力机制中引入上下文信息,以此提取上下文中重要的隐喻语义特征;其次在句法依存树上使用图卷积网络提取隐喻句中的句法结构信息.针对第2个问题,使用双层注意力机制,分别聚焦于单词和句子层面中对隐喻识别和情感分类有贡献的特征信息.在两类任务6个数据集上的对比实验结果表明,该模型相比基线模型性能均有提升. 展开更多
关键词 隐喻识别 情感分类 多任务学习 RoBERTa 图卷积网络 注意力机制
下载PDF
基于ALBERT-Seq2Seq-Attention模型的数字化档案多标签分类
11
作者 王少阳 成新民 +3 位作者 王瑞琴 陈静雯 周阳 费志高 《湖州师范学院学报》 2024年第2期65-72,共8页
针对现有的数字化档案多标签分类方法存在分类标签之间缺少关联性的问题,提出一种用于档案多标签分类的深层神经网络模型ALBERT-Seq2Seq-Attention.该模型通过ALBERT(A Little BERT)预训练语言模型内部多层双向的Transfomer结构获取进... 针对现有的数字化档案多标签分类方法存在分类标签之间缺少关联性的问题,提出一种用于档案多标签分类的深层神经网络模型ALBERT-Seq2Seq-Attention.该模型通过ALBERT(A Little BERT)预训练语言模型内部多层双向的Transfomer结构获取进行文本特征向量的提取,并获得上下文语义信息;将预训练提取的文本特征作为Seq2Seq-Attention(Sequence to Sequence-Attention)模型的输入序列,构建标签字典以获取多标签间的关联关系.将分类模型在3种数据集上分别进行对比实验,结果表明:模型分类的效果F1值均超过90%.该模型不仅能提高档案文本的多标签分类效果,也能关注标签之间的相关关系. 展开更多
关键词 ALBERT Seq2Seq attention 多标签分类 数字化档案
下载PDF
基于Coordinate Attention和空洞卷积的异物识别
12
作者 王春霖 吴春雷 +1 位作者 李灿伟 朱明飞 《计算机系统应用》 2024年第3期178-186,共9页
在我国工厂的工业化生产中,带式运输机占有重要的地位,但是在其运输物料的过程中,常有木板、金属管、大型金属片等混入物料中,从而对带式运输机的传送带造成损毁,引起巨大的经济损失.为了检测出传送带上的不规则异物,设计了一种新的异... 在我国工厂的工业化生产中,带式运输机占有重要的地位,但是在其运输物料的过程中,常有木板、金属管、大型金属片等混入物料中,从而对带式运输机的传送带造成损毁,引起巨大的经济损失.为了检测出传送带上的不规则异物,设计了一种新的异物检测方法.针对传统异物检测方法中存在的对于图像特征提取能力不足以及网络感受野相对较小的问题,我们提出了一种基于coordinate attention和空洞卷积的单阶段异物识别方法.首先,网络利用coordinate attention机制,使网络更加关注图像的空间信息,并对图像中的重要特征进行了增强,增强了网络的性能;其次,在网络提取多尺度特征的部分,将原网络的静态卷积变为空洞卷积,有效减少了常规卷积造成的信息损失;除此之外,我们还使用了新的损失函数,进一步提高了网络的性能.实验结果证明,我们提出的网络能有效识别出传送带上的异物,较好地完成异物检测任务. 展开更多
关键词 coordinate attention 异物检测 空洞卷积 损失函数 目标识别
下载PDF
融合MacBERT和Talking⁃Heads Attention实体关系联合抽取模型
13
作者 王春亮 姚洁仪 李昭 《现代电子技术》 2024年第5期127-131,共5页
针对现有的医学文本关系抽取任务模型在训练过程中存在语义理解能力不足,可能导致关系抽取的效果不尽人意的问题,文中提出一种融合MacBERT和Talking⁃Heads Attention的实体关系联合抽取模型。该模型首先利用MacBERT语言模型来获取动态... 针对现有的医学文本关系抽取任务模型在训练过程中存在语义理解能力不足,可能导致关系抽取的效果不尽人意的问题,文中提出一种融合MacBERT和Talking⁃Heads Attention的实体关系联合抽取模型。该模型首先利用MacBERT语言模型来获取动态字向量表达,MacBERT作为改进的BERT模型,能够减少预训练和微调阶段之间的差异,从而提高模型的泛化能力;然后,将这些动态字向量表达输入到双向门控循环单元(BiGRU)中,以便提取文本的上下文特征。BiGRU是一种改进的循环神经网络(RNN),具有更好的长期依赖捕获能力。在获取文本上下文特征之后,使用Talking⁃Heads Attention来获取全局特征。Talking⁃Heads Attention是一种自注意力机制,可以捕获文本中不同位置之间的关系,从而提高关系抽取的准确性。实验结果表明,与实体关系联合抽取模型GRTE相比,该模型F1值提升1%,precision值提升0.4%,recall值提升1.5%。 展开更多
关键词 MacBERT BiGRU 关系抽取 医学文本 Talking⁃Heads attention 深度学习 全局特征 神经网络
下载PDF
Image Inpainting Technique Incorporating Edge Prior and Attention Mechanism
14
作者 Jinxian Bai Yao Fan +1 位作者 Zhiwei Zhao Lizhi Zheng 《Computers, Materials & Continua》 SCIE EI 2024年第1期999-1025,共27页
Recently,deep learning-based image inpainting methods have made great strides in reconstructing damaged regions.However,these methods often struggle to produce satisfactory results when dealing with missing images wit... Recently,deep learning-based image inpainting methods have made great strides in reconstructing damaged regions.However,these methods often struggle to produce satisfactory results when dealing with missing images with large holes,leading to distortions in the structure and blurring of textures.To address these problems,we combine the advantages of transformers and convolutions to propose an image inpainting method that incorporates edge priors and attention mechanisms.The proposed method aims to improve the results of inpainting large holes in images by enhancing the accuracy of structure restoration and the ability to recover texture details.This method divides the inpainting task into two phases:edge prediction and image inpainting.Specifically,in the edge prediction phase,a transformer architecture is designed to combine axial attention with standard self-attention.This design enhances the extraction capability of global structural features and location awareness.It also balances the complexity of self-attention operations,resulting in accurate prediction of the edge structure in the defective region.In the image inpainting phase,a multi-scale fusion attention module is introduced.This module makes full use of multi-level distant features and enhances local pixel continuity,thereby significantly improving the quality of image inpainting.To evaluate the performance of our method.comparative experiments are conducted on several datasets,including CelebA,Places2,and Facade.Quantitative experiments show that our method outperforms the other mainstream methods.Specifically,it improves Peak Signal-to-Noise Ratio(PSNR)and Structure Similarity Index Measure(SSIM)by 1.141~3.234 db and 0.083~0.235,respectively.Moreover,it reduces Learning Perceptual Image Patch Similarity(LPIPS)and Mean Absolute Error(MAE)by 0.0347~0.1753 and 0.0104~0.0402,respectively.Qualitative experiments reveal that our method excels at reconstructing images with complete structural information and clear texture details.Furthermore,our model exhibits impressive performance in terms of the number of parameters,memory cost,and testing time. 展开更多
关键词 Image inpainting TRANSFORMER edge prior axial attention multi-scale fusion attention
下载PDF
基于Attention-UNet网络的速度模型构建方法研究
15
作者 孙德辉 王云专 王莉利 《物探化探计算技术》 CAS 2024年第1期1-10,共10页
随着油气资源的不断勘探开发,相对易开采的油气矿逐渐建成,地震勘探的研究重点也向地下更深、构造更复杂的区域转移。目前,传统的地震速度建模方法在稳定性、准确性和计算效率方面都面临挑战。因此,笔者利用将地震数据映射到速度模型的... 随着油气资源的不断勘探开发,相对易开采的油气矿逐渐建成,地震勘探的研究重点也向地下更深、构造更复杂的区域转移。目前,传统的地震速度建模方法在稳定性、准确性和计算效率方面都面临挑战。因此,笔者利用将地震数据映射到速度模型的思路,提出了一种基于Attention-UNet网络的深度学习速度建模方法。采用的这种方法利用有限差分正演得到反射波形数据,将反射波形数据和对应的速度模型(标签)对作为Attention-UNet网络的输入,建立地震数据和速度模型之间的映射关系。网络训练后对新输入的地震数据进行速度模型的估计。数值实验结果表明,与传统的FWI相比,笔者提出的方法表现出良好的性能;基于Attention-UNet网络模型训练完成后,不需要经过大量的计算,就可以快速执行与训练集中速度结构相似的地下结构的速度建模,这比传统方法计算效率更高。该方法在建立大量速度模型时具有很好的推广价值。 展开更多
关键词 速度建模 注意力机制 UNet 全波形反演
下载PDF
基于Unet+Attention的胸部CT影像支气管分割算法
16
作者 张子明 周庆华 +1 位作者 薛洪省 覃文军 《中国生物医学工程学报》 CAS CSCD 2024年第1期60-69,共10页
目前肺气管分割中,由于CT图像灰度分布复杂,分割目标像素近似,易造成过分割;而且肺气管像素较少,难以得到更多目标特征,造成细小肺气管容易被忽略。针对这些难点,本研究提出结合Unet网络和注意力机制的肺气管分割算法,注意力机制使用的... 目前肺气管分割中,由于CT图像灰度分布复杂,分割目标像素近似,易造成过分割;而且肺气管像素较少,难以得到更多目标特征,造成细小肺气管容易被忽略。针对这些难点,本研究提出结合Unet网络和注意力机制的肺气管分割算法,注意力机制使用的是关注通道域和空间域的卷积块注意力模型(CBAM),该模型提高了气管特征权重。在损失函数方面,针对原始数据中正负样本失衡的问题,引入focal loss损失函数,该函数对标准交叉熵损失函数进行了改进,使难分类样本在训练过程中得到更多关注;最后通过八连通域判断将孤立点去除,保留较大的几个连通域,即最后的肺气管部分。选用由合作医院提供的24组CT影像和43组CTA影像,共计26157张切片图像作为数据集,进行分割实验。结果表明,分割准确率能够达到0.86,过分割率和欠分割率均值为0.28和0.39。经过注意力模块和损失函数的消融实验,在改进前的准确率、过分割率和欠分割率分别为0.81、0.30、0.40,可见其分割效果均不如Unet+Attention方法。与其他常用方法在相同条件下进行比较后,在保证过分割率和欠分割率不变的情况下,所提出的算法得到了最高的准确率,较好地解决了细小气管分割不准确的问题。 展开更多
关键词 医学图像分割 肺气管 Unet 注意力机制 focal loss
下载PDF
基于Attention-BiTCN的网络入侵检测方法
17
作者 孙红哲 王坚 +1 位作者 王鹏 安雨龙 《信息网络安全》 CSCD 2024年第2期309-318,共10页
为解决网络入侵检测领域多分类准确率不高的问题,文章根据网络流量数据具有时序特征的特点,提出一种基于注意力机制和双向时间卷积神经网络(BiDirectional Temporal Convolutional Network,BiTCN)的网络入侵检测模型。首先,该模型对数... 为解决网络入侵检测领域多分类准确率不高的问题,文章根据网络流量数据具有时序特征的特点,提出一种基于注意力机制和双向时间卷积神经网络(BiDirectional Temporal Convolutional Network,BiTCN)的网络入侵检测模型。首先,该模型对数据集进行独热编码和归一化处置等预处理,解决网络流量数据离散性强和标度不统一的问题;其次,将预处理好的数据经双向滑窗法生成双向序列,并同步输入Attention-Bi TCN模型中;然后,提取双向时序特征并通过加性方式融合,得到时序信息被增强后的融合特征;最后,使用Softmax函数对融合特征进行多种攻击行为检测识别。文章所提模型在NSL-KDD和UNSW-NB15数据集上进行实验验证,多分类准确率分别达到99.70%和84.07%,优于传统网络入侵检测算法,且比其他深度学习模型在检测性能上有显著提升。 展开更多
关键词 入侵检测 注意力机制 BiTCN 双向滑窗法 融合特征
下载PDF
基于Attention-LSTM-XGBoost的电极移动速度影响放电参数预测分析
18
作者 何秀思 阮方鸣 +2 位作者 徐愷 尹兰 王文利 《电波科学学报》 CSCD 2024年第2期287-295,共9页
基于具有Attention机制的长短期记忆(attention long short-term memory,Attention-LSTM)神经网络模型,设计了一种由Attention-LSTM神经网络与极端的梯度增强(extreme gradient boosting,XGBoost)法共同组成的变权组合模型,用以分析预... 基于具有Attention机制的长短期记忆(attention long short-term memory,Attention-LSTM)神经网络模型,设计了一种由Attention-LSTM神经网络与极端的梯度增强(extreme gradient boosting,XGBoost)法共同组成的变权组合模型,用以分析预测静电放电过程中电极移动速度对放电参数造成的影响。该组合模型充分利用静电放电参数的时序特性,并采用Attention机制突出对放电参数预测起到关键作用的输入特征。首先基于由新型电极移动速度效应测试仪的实验结果提供的原始实验数据,采用分箱法对其进行预处理得到新的实验数据;然后将得到的新实验数据集作为两种模型的输入数据,分开训练Attention-LSTM模型和XGBoost模型,求解出各自模型的预测结果及误差;最后利用误差倒数法,重新计算出两种模型预测结果的占比权重,并根据计算的权重求解出最终预测结果。预测结果表明:与Attention-LSTM神经网络模型、XGBoost模型、Attention-LSTMXGBoost定权组合模型相比,本文构建的Attention-LSTM-XGBoost变权组合模型,评估指标中的决定系数分别提升了5.22%、9.11%、3.13%。本文提出的变权组合模型在预测精度以及算法鲁棒性上均优于其他模型,有益于对小间隙静电放电参数变化趋势和规律的探寻。 展开更多
关键词 静电放电 电极移动速度 attention机制 长短期记忆(LSTM)神经网络 误差倒数法
下载PDF
Learning Epipolar Line Window Attention for Stereo Image Super-Resolution Reconstruction
19
作者 Xue Li Hongying Zhang +1 位作者 Zixun Ye Xiaoru 《Computers, Materials & Continua》 SCIE EI 2024年第2期2847-2864,共18页
Transformer-based stereo image super-resolution reconstruction(Stereo SR)methods have significantly improved image quality.However,existing methods have deficiencies in paying attention to detailed features and do not... Transformer-based stereo image super-resolution reconstruction(Stereo SR)methods have significantly improved image quality.However,existing methods have deficiencies in paying attention to detailed features and do not consider the offset of pixels along the epipolar lines in complementary views when integrating stereo information.To address these challenges,this paper introduces a novel epipolar line window attention stereo image super-resolution network(EWASSR).For detail feature restoration,we design a feature extractor based on Transformer and convolutional neural network(CNN),which consists of(shifted)window-based self-attention((S)W-MSA)and feature distillation and enhancement blocks(FDEB).This combination effectively solves the problem of global image perception and local feature attention and captures more discriminative high-frequency features of the image.Furthermore,to address the problem of offset of complementary pixels in stereo images,we propose an epipolar line window attention(EWA)mechanism,which divides windows along the epipolar direction to promote efficient matching of shifted pixels,even in pixel smooth areas.More accurate pixel matching can be achieved using adjacent pixels in the window as a reference.Extensive experiments demonstrate that our EWASSR can reconstruct more realistic detailed features.Comparative quantitative results show that in the experimental results of our EWASSR on the Middlebury and Flickr1024 data sets for 2×SR,compared with the recent network,the Peak signal-to-noise ratio(PSNR)increased by 0.37 dB and 0.34 dB,respectively. 展开更多
关键词 Stereo SR epipolar line window attention feature distillation
下载PDF
The Short-Term Prediction ofWind Power Based on the Convolutional Graph Attention Deep Neural Network
20
作者 Fan Xiao Xiong Ping +4 位作者 Yeyang Li Yusen Xu Yiqun Kang Dan Liu Nianming Zhang 《Energy Engineering》 EI 2024年第2期359-376,共18页
The fluctuation of wind power affects the operating safety and power consumption of the electric power grid and restricts the grid connection of wind power on a large scale.Therefore,wind power forecasting plays a key... The fluctuation of wind power affects the operating safety and power consumption of the electric power grid and restricts the grid connection of wind power on a large scale.Therefore,wind power forecasting plays a key role in improving the safety and economic benefits of the power grid.This paper proposes a wind power predicting method based on a convolutional graph attention deep neural network with multi-wind farm data.Based on the graph attention network and attention mechanism,the method extracts spatial-temporal characteristics from the data of multiple wind farms.Then,combined with a deep neural network,a convolutional graph attention deep neural network model is constructed.Finally,the model is trained with the quantile regression loss function to achieve the wind power deterministic and probabilistic prediction based on multi-wind farm spatial-temporal data.A wind power dataset in the U.S.is taken as an example to demonstrate the efficacy of the proposed model.Compared with the selected baseline methods,the proposed model achieves the best prediction performance.The point prediction errors(i.e.,root mean square error(RMSE)and normalized mean absolute percentage error(NMAPE))are 0.304 MW and 1.177%,respectively.And the comprehensive performance of probabilistic prediction(i.e.,con-tinuously ranked probability score(CRPS))is 0.580.Thus,the significance of multi-wind farm data and spatial-temporal feature extraction module is self-evident. 展开更多
关键词 Format wind power prediction deep neural network graph attention network attention mechanism quantile regression
下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部