期刊文献+
共找到1,633篇文章
< 1 2 82 >
每页显示 20 50 100
Solving Arithmetic Word Problems of Entailing Deep Implicit Relations by Qualia Syntax-Semantic Model
1
作者 Hao Meng Xinguo Yu +3 位作者 Bin He Litian Huang Liang Xue Zongyou Qiu 《Computers, Materials & Continua》 SCIE EI 2023年第10期541-555,共15页
Solving arithmetic word problems that entail deep implicit relations is still a challenging problem.However,significant progress has been made in solving Arithmetic Word Problems(AWP)over the past six decades.This pap... Solving arithmetic word problems that entail deep implicit relations is still a challenging problem.However,significant progress has been made in solving Arithmetic Word Problems(AWP)over the past six decades.This paper proposes to discover deep implicit relations by qualia inference to solve Arithmetic Word Problems entailing Deep Implicit Relations(DIR-AWP),such as entailing commonsense or subject-domain knowledge involved in the problem-solving process.This paper proposes to take three steps to solve DIR-AWPs,in which the first three steps are used to conduct the qualia inference process.The first step uses the prepared set of qualia-quantity models to identify qualia scenes from the explicit relations extracted by the Syntax-Semantic(S2)method from the given problem.The second step adds missing entities and deep implicit relations in order using the identified qualia scenes and the qualia-quantity models,respectively.The third step distills the relations for solving the given problem by pruning the spare branches of the qualia dependency graph of all the acquired relations.The research contributes to the field by presenting a comprehensive approach combining explicit and implicit knowledge to enhance reasoning abilities.The experimental results on Math23K demonstrate hat the proposed algorithm is superior to the baseline algorithms in solving AWPs requiring deep implicit relations. 展开更多
关键词 Arithmetic word problem implicit quantity relations qualia syntax-semantic model
下载PDF
基于图像Bag-of-Words模型的无载体信息隐藏 被引量:42
2
作者 周志立 曹燚 孙星明 《应用科学学报》 CAS CSCD 北大核心 2016年第5期527-536,共10页
介绍一种基于bag-of-words(BOW)模型的无载体信息隐藏方法.该方法使用BOW模型提取图像的视觉关键词(visual words,VW)以表达待隐藏的文本信息,从而实现文本信息在图像中的隐藏.首先使用BOW模型提取图像集中每幅图像的VW,构建文本信息的... 介绍一种基于bag-of-words(BOW)模型的无载体信息隐藏方法.该方法使用BOW模型提取图像的视觉关键词(visual words,VW)以表达待隐藏的文本信息,从而实现文本信息在图像中的隐藏.首先使用BOW模型提取图像集中每幅图像的VW,构建文本信息的关键词和VW的映射关系库;然后把每幅图像分为若干子图像,统计每一幅子图像的VW频数直方图,选择频数最高的VW表示该子图像;最后根据构建的文本关键词和子图像VW的映射关系库,搜索出与待隐藏文本信息存在映射关系的子图像序列,将含有这些子图像的图像作为含密图像进行传递.实验结果和分析表明,该隐藏算法在抗隐写分析、鲁棒性和安全性方面均有良好的表现. 展开更多
关键词 无载体信息隐藏 BAG of words模型 视觉词汇 图像搜索
下载PDF
n-words模型下Hesse稀疏表示的图像检索算法 被引量:2
3
作者 王瑞霞 彭国华 《电子与信息学报》 EI CSCD 北大核心 2016年第5期1115-1122,共8页
论文针对视觉词袋(BOVW)模型放弃图像空间结构的缺点,提出一种基于Hesse稀疏编码的图像检索算法。首先,建立n-words模型,获得图像局部特征表示。n-words模型由一系列连续视觉词获得,是图像特征的一种高级描述。该文从n=1到n=5进行试验,... 论文针对视觉词袋(BOVW)模型放弃图像空间结构的缺点,提出一种基于Hesse稀疏编码的图像检索算法。首先,建立n-words模型,获得图像局部特征表示。n-words模型由一系列连续视觉词获得,是图像特征的一种高级描述。该文从n=1到n=5进行试验,寻找最恰当的n值;其次,将二阶Hesse能量函数融入标准稀疏编码的目标函数,得到Hesse稀疏编码公式;最后,以获得的n-words序列作为编码特征,利用特征符号搜索算法求解最优Hesse系数,计算相似度,返回检索结果。实验在两类数据集上进行,与BOVW模型和已有的算法相比,新算法极大地提高了图像检索的准确率。 展开更多
关键词 图像检索 稀疏编码 视觉词袋模型 n-words模型 Hesse能量函数
下载PDF
基于Bag of Words的干果图像分类研究 被引量:1
4
作者 施明登 周鹏 白铁成 《安徽农业科学》 CAS 2014年第29期10381-10383,共3页
针对干果图像信息量大、分类精度低和耗时多的特点,提出利用Bag of Words模型提取图片的代表特征,并采用朴素贝叶斯分类器指导特征矩阵分类。结果表明,图像分类精度能达到80%,分类处理时间约为2 s。通过增加学习样本来进一步提高分类精... 针对干果图像信息量大、分类精度低和耗时多的特点,提出利用Bag of Words模型提取图片的代表特征,并采用朴素贝叶斯分类器指导特征矩阵分类。结果表明,图像分类精度能达到80%,分类处理时间约为2 s。通过增加学习样本来进一步提高分类精度,将Bag of Words应用于干果图像识别和分类是可行的。 展开更多
关键词 图像分类 词袋模型 朴素贝叶斯分类器
下载PDF
Chinese New-coined Words Management in P.R.C
5
作者 赵鹏 柳轶群 《海外英语》 2014年第7X期257-258,260,共3页
The paper relies on the language management model. First of all, the thousands of new words coined each year in China can be roughly divided into eight groups. Besides, it exemplifies the problems that have occurred i... The paper relies on the language management model. First of all, the thousands of new words coined each year in China can be roughly divided into eight groups. Besides, it exemplifies the problems that have occurred in communication. Last but not least, to reflect language management acts/processes, it reports the efforts the language authorities and governmental bodies have made in the struggle to solve the problems. It attempts to remind the language management agencies of the point that for the use of new words, the language users are always right as long as communication flows without problems in understanding. 展开更多
关键词 LANGUAGE MANAGEMENT model new-coined words purify
下载PDF
A Study on Translation of Chinese Culture-loaded Words from the Per-spective of Relevance Theory
6
作者 高芳卉 《海外英语》 2018年第11期151-153,共3页
The essay tends to analyze the translation of Chinese Culture-loaded words from the perspective of Relevance Theory.The theory gains its prominence by studying the translation process and transcending the conflicts be... The essay tends to analyze the translation of Chinese Culture-loaded words from the perspective of Relevance Theory.The theory gains its prominence by studying the translation process and transcending the conflicts between literal and free translation. It incorporates recent work in cognitive linguistics, with ostensive-inference as its key model. Under the influence of Relevance theory, the translation of culture-loaded words is reader-oriented. Translators are obliged to help target readers to establish new assumptions to achieve equivalent response. 展开更多
关键词 Relevance Theory ostensive-inferential model culture-loaded words contextual effects
下载PDF
Investigating the Psychometric Impact of Negative Worded Items in Reading Comprehension Passages with a 3PL Cross-Classified Testlet Model
7
作者 Yong Luo Junhui Liu 《Journal of International Education and Practice》 2019年第1期47-59,共13页
Negative worded(NW)items used in psychological instruments have been studied with the bifactor model to investigate whether the NW items form a secondary factor due to negative wording orthogonal to the measured laten... Negative worded(NW)items used in psychological instruments have been studied with the bifactor model to investigate whether the NW items form a secondary factor due to negative wording orthogonal to the measured latent construct,a validation procedure which checks whether NW items form a source of construct irrelevant variance(CIV)and hence constitute a validity threat.In the context of educational testing,however,no such validation attempts have been made.In this study,we studied the psychometric impact of NW items in an English proficiency reading comprehension test using a modeling approach similar to the bifactor model,namely the three-parameter logistic cross-classified testlet response theory(3PL CCTRT)model,to account for both guessing and possible local item dependence due to passage effect in the data set.The findings indicate that modeling the NW items with a separate factor leads to noticeable improvement in model fit,and the factor variance is marginal but nonzero.However,item and ability parameter estimates are highly similar between the 3PL CCTRT model and other models that do not model the NW items.It is concluded that the NW items introduce CIV into the data,but its magnitude is too small to change item and person ability parameter estimates to an extent of practical significance. 展开更多
关键词 Negative wording Bifactor model Cross-classified testlet model VALIDATION
下载PDF
Vari-gram language model based on word clustering
8
作者 袁里驰 《Journal of Central South University》 SCIE EI CAS 2012年第4期1057-1062,共6页
Category-based statistic language model is an important method to solve the problem of sparse data.But there are two bottlenecks:1) The problem of word clustering.It is hard to find a suitable clustering method with g... Category-based statistic language model is an important method to solve the problem of sparse data.But there are two bottlenecks:1) The problem of word clustering.It is hard to find a suitable clustering method with good performance and less computation.2) Class-based method always loses the prediction ability to adapt the text in different domains.In order to solve above problems,a definition of word similarity by utilizing mutual information was presented.Based on word similarity,the definition of word set similarity was given.Experiments show that word clustering algorithm based on similarity is better than conventional greedy clustering method in speed and performance,and the perplexity is reduced from 283 to 218.At the same time,an absolute weighted difference method was presented and was used to construct vari-gram language model which has good prediction ability.The perplexity of vari-gram model is reduced from 234.65 to 219.14 on Chinese corpora,and is reduced from 195.56 to 184.25 on English corpora compared with category-based model. 展开更多
关键词 统计语言模型 词聚类 词语相似度 聚类分析方法 预测能力 聚类方法 聚类算法 相似性
下载PDF
Research on high-performance English translation based on topic model
9
作者 Yumin Shen Hongyu Guo 《Digital Communications and Networks》 SCIE CSCD 2023年第2期505-511,共7页
Retelling extraction is an important branch of Natural Language Processing(NLP),and high-quality retelling resources are very helpful to improve the performance of machine translation.However,traditional methods based... Retelling extraction is an important branch of Natural Language Processing(NLP),and high-quality retelling resources are very helpful to improve the performance of machine translation.However,traditional methods based on the bilingual parallel corpus often ignore the document background in the process of retelling acquisition and application.In order to solve this problem,we introduce topic model information into the translation mode and propose a topic-based statistical machine translation method to improve the translation performance.In this method,Probabilistic Latent Semantic Analysis(PLSA)is used to obtains the co-occurrence relationship between words and documents by the hybrid matrix decomposition.Then we design a decoder to simplify the decoding process.Experiments show that the proposed method can effectively improve the accuracy of translation. 展开更多
关键词 Machine translation Topic model Statistical machine translation Bilingual word vector RETELLING
下载PDF
An Effective Machine-Learning Based Feature Extraction/Recognition Model for Fetal Heart Defect Detection from 2D Ultrasonic Imageries
10
作者 Bingzheng Wu Peizhong Liu +3 位作者 Huiling Wu Shunlan Liu Shaozheng He Guorong Lv 《Computer Modeling in Engineering & Sciences》 SCIE EI 2023年第2期1069-1089,共21页
Congenital heart defect,accounting for about 30%of congenital defects,is the most common one.Data shows that congenital heart defects have seriously affected the birth rate of healthy newborns.In Fetal andNeonatal Car... Congenital heart defect,accounting for about 30%of congenital defects,is the most common one.Data shows that congenital heart defects have seriously affected the birth rate of healthy newborns.In Fetal andNeonatal Cardiology,medical imaging technology(2D ultrasonic,MRI)has been proved to be helpful to detect congenital defects of the fetal heart and assists sonographers in prenatal diagnosis.It is a highly complex task to recognize 2D fetal heart ultrasonic standard plane(FHUSP)manually.Compared withmanual identification,automatic identification through artificial intelligence can save a lot of time,ensure the efficiency of diagnosis,and improve the accuracy of diagnosis.In this study,a feature extraction method based on texture features(Local Binary Pattern LBP and Histogram of Oriented Gradient HOG)and combined with Bag of Words(BOW)model is carried out,and then feature fusion is performed.Finally,it adopts Support VectorMachine(SVM)to realize automatic recognition and classification of FHUSP.The data includes 788 standard plane data sets and 448 normal and abnormal plane data sets.Compared with some other methods and the single method model,the classification accuracy of our model has been obviously improved,with the highest accuracy reaching 87.35%.Similarly,we also verify the performance of the model in normal and abnormal planes,and the average accuracy in classifying abnormal and normal planes is 84.92%.The experimental results show that thismethod can effectively classify and predict different FHUSP and can provide certain assistance for sonographers to diagnose fetal congenital heart disease. 展开更多
关键词 Congenital heart defect fetal heart ultrasonic standard plane image recognition and classification machine learning bag of words model feature fusion
下载PDF
Word Embeddings and Semantic Spaces in Natural Language Processing
11
作者 Peter J. Worth 《International Journal of Intelligence Science》 2023年第1期1-21,共21页
One of the critical hurdles, and breakthroughs, in the field of Natural Language Processing (NLP) in the last two decades has been the development of techniques for text representation that solves the so-called curse ... One of the critical hurdles, and breakthroughs, in the field of Natural Language Processing (NLP) in the last two decades has been the development of techniques for text representation that solves the so-called curse of dimensionality, a problem which plagues NLP in general given that the feature set for learning starts as a function of the size of the language in question, upwards of hundreds of thousands of terms typically. As such, much of the research and development in NLP in the last two decades has been in finding and optimizing solutions to this problem, to feature selection in NLP effectively. This paper looks at the development of these various techniques, leveraging a variety of statistical methods which rest on linguistic theories that were advanced in the middle of the last century, namely the distributional hypothesis which suggests that words that are found in similar contexts generally have similar meanings. In this survey paper we look at the development of some of the most popular of these techniques from a mathematical as well as data structure perspective, from Latent Semantic Analysis to Vector Space Models to their more modern variants which are typically referred to as word embeddings. In this review of algoriths such as Word2Vec, GloVe, ELMo and BERT, we explore the idea of semantic spaces more generally beyond applicability to NLP. 展开更多
关键词 Natural Language Processing Vector Space models Semantic Spaces word Embeddings Representation Learning Text Vectorization Machine Learning Deep Learning
下载PDF
Naxi-English Bilingual Word Alignment Based on Language Characteristics and Log-Linear Model
12
作者 Yu Zhengtao Xian Yantuan +2 位作者 Tian Wei Guo Jianyi Zhang Tao 《China Communications》 SCIE CSCD 2012年第3期78-86,共9页
We propose a method that can achieve the Naxi-English bilingual word automatic alignment based on a log-linear model.This method defines the different Naxi-English structural feature functions,which are English-Naxi i... We propose a method that can achieve the Naxi-English bilingual word automatic alignment based on a log-linear model.This method defines the different Naxi-English structural feature functions,which are English-Naxi interval switching function and Naxi-English bilingual word position transformation function.With the manually labeled Naxi-English words alignment corpus,the parameters of the model are trained by using the minimum error,thus Naxi-English bilingual word alignment is achieved automatically.Experiments are conducted with IBM Model 3 as a benchmark,and the Naxi language constraints are introduced.The final experiment results show that the proposed alignment method achieves very good results:the introduction of the language characteristic function can effectively improve the accuracy of the Naxi-English Bilingual Word Alignment. 展开更多
关键词 对数线性模型 词对齐 双语 语言 纳西族 自动对齐 结构特征 转换功能
下载PDF
自然语言处理领域中的词嵌入方法综述
13
作者 曾骏 王子威 +2 位作者 于扬 文俊浩 高旻 《计算机科学与探索》 CSCD 北大核心 2024年第1期24-43,共20页
词嵌入作为自然语言处理任务的第一步,其目的是将输入的自然语言文本转换为模型可以处理的数值向量,即词向量,也称词的分布式表示。词向量作为自然语言处理任务的根基,是完成一切自然语言处理任务的前提。然而,国内外针对词嵌入方法的... 词嵌入作为自然语言处理任务的第一步,其目的是将输入的自然语言文本转换为模型可以处理的数值向量,即词向量,也称词的分布式表示。词向量作为自然语言处理任务的根基,是完成一切自然语言处理任务的前提。然而,国内外针对词嵌入方法的综述文献大多只关注于不同词嵌入方法本身的技术路线,而未能将词嵌入的前置分词方法以及词嵌入方法完整的演变趋势进行分析与概述。以word2vec模型和Transformer模型作为划分点,从生成的词向量是否能够动态地改变其内隐的语义信息来适配输入句子的整体语义这一角度,将词嵌入方法划分为静态词嵌入方法和动态词嵌入方法,并对此展开讨论。同时,针对词嵌入中的分词方法,包括整词切分和子词切分,进行了对比和分析;针对训练词向量所使用的语言模型,从概率语言模型到神经概率语言模型再到如今的深度上下文语言模型的演化,进行了详细列举和阐述;针对预训练语言模型时使用的训练策略进行了总结和探讨。最后,总结词向量质量的评估方法,分析词嵌入方法的当前现状并对其未来发展方向进行展望。 展开更多
关键词 词向量 词嵌入方法 自然语言处理 语言模型 分词 词向量评估
下载PDF
主题方面共享的领域主题层次模型
14
作者 万常选 张奕韬 +3 位作者 刘德喜 刘喜平 廖国琼 万齐智 《软件学报》 EI CSCD 北大核心 2024年第4期1790-1818,共29页
层次主题模型是构建主题层次的重要工具.现有的层次主题模型大多通过在主题模型中引入nCRP构造方法,为文档主题提供树形结构的先验分布,但无法生成具有明确领域涵义的主题层次结构,即领域主题层次.同时,领域主题不仅存在层次关系,而且... 层次主题模型是构建主题层次的重要工具.现有的层次主题模型大多通过在主题模型中引入nCRP构造方法,为文档主题提供树形结构的先验分布,但无法生成具有明确领域涵义的主题层次结构,即领域主题层次.同时,领域主题不仅存在层次关系,而且不同父主题下的子主题之间还存在子领域方面共享的关联关系,在现有主题关系研究中没有合适的模型来生成这种领域主题层次.为了从领域文本中自动、有效地挖掘出领域主题的层次关系和关联关系,在4个方面进行创新研究.首先,通过主题共享机制改进nCRP构造方法,提出nCRP+层次构造方法,为主题模型中的主题提供具有分层主题方面共享的树形先验分布;其次,结合nCRP+和HDP模型构建重分层的Dirichlet过程,提出rHDP(reallocated hierarchical Dirichlet processes)层次主题模型;第三,结合领域分类信息、词语语义和主题词的领域代表性,定义领域知识,包括基于投票机制的领域隶属度、词语与领域主题的语义相关度和层次化的主题-词语贡献度;最后,通过领域知识改进rHDP主题模型中领域主题和主题词的分配过程,提出结合领域知识的层次主题模型rHDP_DK(rHDP with domain knowledge),并改进采样过程.实验结果表明,基于nCRP+的层次主题模型在评价指标方面均优于基于nCRP的层次主题模型(hLDA,nHDP)和神经主题模型(TSNTM);通过rHDP_DK模型生成的主题层次结构具有领域主题层次清晰、关联子主题的主题词领域差异明确的特点.此外,该模型将为领域主题层次提供一个通用的自动挖掘框架. 展开更多
关键词 层次主题模型 领域分类信息 词语语义 主题关联关系 层次化的采样过程 领域主题层次
下载PDF
英语单词学习推荐模型在教学改革中的应用研究
15
作者 胡二娟 刘小强 《计算机应用文摘》 2024年第8期5-7,共3页
在人工智能技术的背景下,文章主要探讨了英语单词学习的推荐方法。其中,建立了1个基于人工智能的英语单词学习推荐模型,该模型利用用户的学习行为数据和单词特征进行训练,旨在为用户提供个性化的单词推荐。实验验证结果显示,相较于传统... 在人工智能技术的背景下,文章主要探讨了英语单词学习的推荐方法。其中,建立了1个基于人工智能的英语单词学习推荐模型,该模型利用用户的学习行为数据和单词特征进行训练,旨在为用户提供个性化的单词推荐。实验验证结果显示,相较于传统方法,该模型具有更高的准确性和可靠性,能够有效提升英语单词学习的效果。 展开更多
关键词 人工智能 英语单词 推荐模型 模型构建
下载PDF
基于BERT-BiLSTM-CRF模型的畜禽疫病文本分词研究
16
作者 余礼根 郭晓利 +3 位作者 赵红涛 杨淦 张俊 李奇峰 《农业机械学报》 EI CAS CSCD 北大核心 2024年第2期287-294,共8页
针对畜禽疫病文本语料匮乏、文本内包含大量疫病名称及短语等未登录词问题,提出了一种结合词典匹配的BERT-BiLSTM-CRF畜禽疫病文本分词模型。以羊疫病为研究对象,构建了常见疫病文本数据集,将其与通用语料PKU结合,利用BERT(Bidirectiona... 针对畜禽疫病文本语料匮乏、文本内包含大量疫病名称及短语等未登录词问题,提出了一种结合词典匹配的BERT-BiLSTM-CRF畜禽疫病文本分词模型。以羊疫病为研究对象,构建了常见疫病文本数据集,将其与通用语料PKU结合,利用BERT(Bidirectional encoder representation from transformers)预训练语言模型进行文本向量化表示;通过双向长短时记忆网络(Bidirectional long short-term memory network,BiLSTM)获取上下文语义特征;由条件随机场(Conditional random field,CRF)输出全局最优标签序列。基于此,在CRF层后加入畜禽疫病领域词典进行分词匹配修正,减少在分词过程中出现的疫病名称及短语等造成的歧义切分,进一步提高了分词准确率。实验结果表明,结合词典匹配的BERT-BiLSTM-CRF模型在羊常见疫病文本数据集上的F1值为96.38%,与jieba分词器、BiLSTM-Softmax模型、BiLSTM-CRF模型、未结合词典匹配的本文模型相比,分别提升11.01、10.62、8.3、0.72个百分点,验证了方法的有效性。与单一语料相比,通用语料PKU和羊常见疫病文本数据集结合的混合语料,能够同时对畜禽疫病专业术语及疫病文本中常用词进行准确切分,在通用语料及疫病文本数据集上F1值都达到95%以上,具有较好的模型泛化能力。该方法可用于畜禽疫病文本分词。 展开更多
关键词 畜禽疫病 文本分词 预训练语言模型 双向长短时记忆网络 条件随机场
下载PDF
基于对抗训练的事件要素识别方法
17
作者 廖涛 沈文龙 +1 位作者 张顺香 马文祥 《计算机工程与设计》 北大核心 2024年第2期540-545,共6页
针对目前大多数事件要素识别模型未考虑词级别的语义信息,及模型鲁棒性不高的问题,提出一种融合词信息和对抗训练的事件要素识别方法。将Bert(bidirectional encode representations from transformers)预训练语言模型生成的字向量与分... 针对目前大多数事件要素识别模型未考虑词级别的语义信息,及模型鲁棒性不高的问题,提出一种融合词信息和对抗训练的事件要素识别方法。将Bert(bidirectional encode representations from transformers)预训练语言模型生成的字向量与分词信息进行融合,在得到的融合向量中添加扰动因子产生对抗样本,将对抗样本与融合向量表示作为编码层的输入;采用BiGRU(bidirectional gating recurrent unit)网络对输入的文本进行编码,丰富文本的上下文语义信息;采用CRF(conditional random field)函数计算完成事件要素的识别任务。在CEC(Chinese emergency corpus)中文突发事件语料库上的实验结果表明,该方法能够取得较好的效果。 展开更多
关键词 事件要素识别 鲁棒性 词信息 对抗训练 预训练语言模型 扰动因子 上下文语义信息
下载PDF
高斯混合模型与文本图卷积网络结合的虚假评论识别算法
18
作者 王星 刘贵娟 陈志豪 《计算机应用》 CSCD 北大核心 2024年第2期360-368,共9页
针对文本图卷积网络(Text GCN)窗口边权阈值策略不足的问题,为了更精准地挖掘相关的词关联结构、提高预测精度,提出一种高斯混合模型(GMM)与Text GCN结合的虚假评论识别算法F-Text GCN。首先,利用GMM分离噪声边权分布的特性,提高虚假评... 针对文本图卷积网络(Text GCN)窗口边权阈值策略不足的问题,为了更精准地挖掘相关的词关联结构、提高预测精度,提出一种高斯混合模型(GMM)与Text GCN结合的虚假评论识别算法F-Text GCN。首先,利用GMM分离噪声边权分布的特性,提高虚假评论在训练数据上相对正常评论数不足的边信号强度;然后,考虑到信源的多样性,综合文档、词汇和评论以及非文本特征构造邻接矩阵;最后,通过Text GCN的谱分解提取邻接矩阵的虚假评论关联结构实施预测。根据国内某大型电商平台采集的126086条实际中文评论数据开展实证研究,实验结果表明,F-Text GCN识别虚假评论的F1值达到82.92%,与预训练表征模型BERT和文本卷积神经网络相比分别提升了10.46%和11.60%,相较于只使用评论文本信源的Text GCN模型F1值提升了2.94%;研究了高仿虚假评论的预测错误率,在支持向量机(SVM)作用后难识别的评论样本上尝试二次识别,F-Text GCN整体预测准确率可达94.71%,相较于Text GCN和SVM,在识别准确率上分别提升了2.91%和14.54%。研究发现,虚假评论的二阶图邻居结构显示出较强的干预消费者决策的词汇,这表明所提算法特别适用于提取用于虚假评论检测的长程词语搭配结构和全局句子特征模式变化的场景。 展开更多
关键词 高斯混合模型 虚假评论识别 文本图卷积神经网络 邻接矩阵 词汇共现网络
下载PDF
基于隐含空间模型降维和LDA模型的学科主题识别研究
19
作者 王婧 武帅 《情报探索》 2024年第2期1-11,共11页
【目的/意义】现有学科研究主题的梳理多为领域专家的定性分析和学科学者的文献梳理,一定程度会由于研究思维的局限性和获取知识的片面性造成学科研究主题误判,为有效避免漏判误判现象的发生,提出识别模型。【方法/过程】首先,运用传统... 【目的/意义】现有学科研究主题的梳理多为领域专家的定性分析和学科学者的文献梳理,一定程度会由于研究思维的局限性和获取知识的片面性造成学科研究主题误判,为有效避免漏判误判现象的发生,提出识别模型。【方法/过程】首先,运用传统LDA模型分析主题特征词;其次,结合上下文语义信息进行中文分词,形成学科主题词库;最后,结合隐含位置聚类算法发现潜在社区,提高主题识别效果。【结果/结论】提出的方法一定程度上优化了主题挖掘算法在识别短文本主题的效果,消除主观意愿。由计算机自行分类并实现科学研究前沿主题的预测,揭示前沿领域的研究热点,为致力于研究前沿学科的新兴学者提供参考价值。 展开更多
关键词 学科主题识别 LDA主题挖掘 图书情报与档案管理学科词库 隐含位置聚类模型 共词网络
下载PDF
使用Visual Basic操纵Microsoft Word对象生成报表文档 被引量:12
20
作者 孔令彦 董蓬勃 +1 位作者 姜青香 刘慧平 《计算机工程与应用》 CSCD 北大核心 2003年第36期115-117,共3页
在使用VisualBasic进行管理信息系统开发中,经常需要操纵MicrosoftWord在文档中特定位置填入从数据库中读取的数据资料来生成所需报表文档,以便于管理或打印输出。文章介绍了MicrosoftWord对象模型,以及在VisuaBasic应用程序中如何利用... 在使用VisualBasic进行管理信息系统开发中,经常需要操纵MicrosoftWord在文档中特定位置填入从数据库中读取的数据资料来生成所需报表文档,以便于管理或打印输出。文章介绍了MicrosoftWord对象模型,以及在VisuaBasic应用程序中如何利用OLE引用MicrosoftWord对象。并且通过实例说明了在VisualBasic中利用MicrosoftWord对象的书签(Bookmark)和查找替换(Find、Replacement)两种方式向模板文档中填入数据实现报表文档的生成和保存的方法。 展开更多
关键词 word VISUALBASIC 报表文档 对象模型 数据库 软件开发 文字处理软件 管理信息系统
下载PDF
上一页 1 2 82 下一页 到第
使用帮助 返回顶部