期刊文献+
共找到32篇文章
< 1 2 >
每页显示 20 50 100
Self-potential inversion based on Attention U-Net deep learning network
1
作者 GUO You-jun CUI Yi-an +3 位作者 CHEN Hang XIE Jing ZHANG Chi LIU Jian-xin 《Journal of Central South University》 SCIE EI CAS CSCD 2024年第9期3156-3167,共12页
Landfill leaks pose a serious threat to environmental health,risking the contamination of both groundwater and soil resources.Accurate investigation of these sites is essential for implementing effective prevention an... Landfill leaks pose a serious threat to environmental health,risking the contamination of both groundwater and soil resources.Accurate investigation of these sites is essential for implementing effective prevention and control measures.The self-potential(SP)stands out for its sensitivity to contamination plumes,offering a solution for monitoring and detecting the movement and seepage of subsurface pollutants.However,traditional SP inversion techniques heavily rely on precise subsurface resistivity information.In this study,we propose the Attention U-Net deep learning network for rapid SP inversion.By incorporating an attention mechanism,this algorithm effectively learns the relationship between array-style SP data and the location and extent of subsurface contaminated sources.We designed a synthetic landfill model with a heterogeneous resistivity structure to assess the performance of Attention U-Net deep learning network.Additionally,we conducted further validation using a laboratory model to assess its practical applicability.The results demonstrate that the algorithm is not solely dependent on resistivity information,enabling effective locating of the source distribution,even in models with intricate subsurface structures.Our work provides a promising tool for SP data processing,enhancing the applicability of this method in the field of near-subsurface environmental monitoring. 展开更多
关键词 SELF-POTENTIAL attention mechanism U-Net deep learning network INVERSION landfill
下载PDF
Semantic Pneumonia Segmentation and Classification for Covid-19 Using Deep Learning Network
2
作者 M.M.Lotfy Hazem M.El-Bakry +4 位作者 M.M.Elgayar Shaker El-Sappagh G.Abdallah M.I A.A.Soliman Kyung Sup Kwak 《Computers, Materials & Continua》 SCIE EI 2022年第10期1141-1158,共18页
Early detection of the Covid-19 disease is essential due to its higher rate of infection affecting tens of millions of people,and its high number of deaths also by 7%.For that purpose,a proposed model of several stage... Early detection of the Covid-19 disease is essential due to its higher rate of infection affecting tens of millions of people,and its high number of deaths also by 7%.For that purpose,a proposed model of several stages was developed.The first stage is optimizing the images using dynamic adaptive histogram equalization,performing a semantic segmentation using DeepLabv3Plus,then augmenting the data by flipping it horizontally,rotating it,then flipping it vertically.The second stage builds a custom convolutional neural network model using several pre-trained ImageNet.Finally,the model compares the pre-trained data to the new output,while repeatedly trimming the best-performing models to reduce complexity and improve memory efficiency.Several experiments were done using different techniques and parameters.Accordingly,the proposed model achieved an average accuracy of 99.6%and an area under the curve of 0.996 in the Covid-19 detection.This paper will discuss how to train a customized intelligent convolutional neural network using various parameters on a set of chest X-rays with an accuracy of 99.6%. 展开更多
关键词 SARS-COV2 COVID-19 PNEUMONIA deep learning network semantic segmentation smart classification
下载PDF
Deep Learning Network for Energy Storage Scheduling in Power Market Environment Short-Term Load Forecasting Model
3
作者 Yunlei Zhang RuifengCao +3 位作者 Danhuang Dong Sha Peng RuoyunDu Xiaomin Xu 《Energy Engineering》 EI 2022年第5期1829-1841,共13页
In the electricity market,fluctuations in real-time prices are unstable,and changes in short-term load are determined by many factors.By studying the timing of charging and discharging,as well as the economic benefits... In the electricity market,fluctuations in real-time prices are unstable,and changes in short-term load are determined by many factors.By studying the timing of charging and discharging,as well as the economic benefits of energy storage in the process of participating in the power market,this paper takes energy storage scheduling as merely one factor affecting short-term power load,which affects short-term load time series along with time-of-use price,holidays,and temperature.A deep learning network is used to predict the short-term load,a convolutional neural network(CNN)is used to extract the features,and a long short-term memory(LSTM)network is used to learn the temporal characteristics of the load value,which can effectively improve prediction accuracy.Taking the load data of a certain region as an example,the CNN-LSTM prediction model is compared with the single LSTM prediction model.The experimental results show that the CNN-LSTM deep learning network with the participation of energy storage in dispatching can have high prediction accuracy for short-term power load forecasting. 展开更多
关键词 Energy storage scheduling short-term load forecasting deep learning network convolutional neural network CNN long and short term memory network LTSM
下载PDF
Prediction of Flash Flood Susceptibility of Hilly Terrain Using Deep Neural Network:A Case Study of Vietnam 被引量:2
4
作者 Huong Thi Thanh Ngo Nguyen Duc Dam +7 位作者 Quynh-Anh Thi Bui Nadhir Al-Ansari Romulus Costache Hang Ha Quynh Duy Bui Sy Hung Mai Indra Prakash Binh Thai Pham 《Computer Modeling in Engineering & Sciences》 SCIE EI 2023年第6期2219-2241,共23页
Flash floods are one of the most dangerous natural disasters,especially in hilly terrain,causing loss of life,property,and infrastructures and sudden disruption of traffic.These types of floods are mostly associated w... Flash floods are one of the most dangerous natural disasters,especially in hilly terrain,causing loss of life,property,and infrastructures and sudden disruption of traffic.These types of floods are mostly associated with landslides and erosion of roads within a short time.Most of Vietnamis hilly and mountainous;thus,the problem due to flash flood is severe and requires systematic studies to correctly identify flood susceptible areas for proper landuse planning and traffic management.In this study,three Machine Learning(ML)methods namely Deep Learning Neural Network(DL),Correlation-based FeatureWeighted Naive Bayes(CFWNB),and Adaboost(AB-CFWNB)were used for the development of flash flood susceptibility maps for hilly road section(115 km length)of National Highway(NH)-6 inHoa Binh province,Vietnam.In the proposedmodels,88 past flash flood events were used together with 14 flash floods affecting topographical and geo-environmental factors.The performance of themodels was evaluated using standard statisticalmeasures including Receiver Operating Characteristic(ROC)Curve,Area Under Curve(AUC)and Root Mean Square Error(RMSE).The results revealed that all the models performed well(AUC>0.80)in predicting flash flood susceptibility zones,but the performance of the DL model is the best(AUC:0.972,RMSE:0.352).Therefore,the DL model can be applied to develop an accurate flash flood susceptibility map of hilly terrain which can be used for proper planning and designing of the highways and other infrastructure facilities besides landuse management of the area. 展开更多
关键词 Flash flood deep learning neural network(DL) machine learning(ML) receiver operating characteristic curve(ROC) VIETNAM
下载PDF
DeepIoT.IDS:Hybrid Deep Learning for Enhancing IoT Network Intrusion Detection 被引量:4
5
作者 Ziadoon K.Maseer Robiah Yusof +3 位作者 Salama A.Mostafa Nazrulazhar Bahaman Omar Musa Bander Ali Saleh Al-rimy 《Computers, Materials & Continua》 SCIE EI 2021年第12期3945-3966,共22页
With an increasing number of services connected to the internet,including cloud computing and Internet of Things(IoT)systems,the prevention of cyberattacks has become more challenging due to the high dimensionality of... With an increasing number of services connected to the internet,including cloud computing and Internet of Things(IoT)systems,the prevention of cyberattacks has become more challenging due to the high dimensionality of the network traffic data and access points.Recently,researchers have suggested deep learning(DL)algorithms to define intrusion features through training empirical data and learning anomaly patterns of attacks.However,due to the high dynamics and imbalanced nature of the data,the existing DL classifiers are not completely effective at distinguishing between abnormal and normal behavior line connections for modern networks.Therefore,it is important to design a self-adaptive model for an intrusion detection system(IDS)to improve the detection of attacks.Consequently,in this paper,a novel hybrid weighted deep belief network(HW-DBN)algorithm is proposed for building an efficient and reliable IDS(DeepIoT.IDS)model to detect existing and novel cyberattacks.The HW-DBN algorithm integrates an improved Gaussian–Bernoulli restricted Boltzmann machine(Deep GB-RBM)feature learning operator with a weighted deep neural networks(WDNN)classifier.The CICIDS2017 dataset is selected to evaluate the DeepIoT.IDS model as it contains multiple types of attacks,complex data patterns,noise values,and imbalanced classes.We have compared the performance of the DeepIoT.IDS model with three recent models.The results show the DeepIoT.IDS model outperforms the three other models by achieving a higher detection accuracy of 99.38%and 99.99%for web attack and bot attack scenarios,respectively.Furthermore,it can detect the occurrence of low-frequency attacks that are undetectable by other models. 展开更多
关键词 Cyberattacks internet of things intrusion detection system deep learning neural network supervised and unsupervised deep learning
下载PDF
Hybrid Deep Learning-Improved BAT Optimization Algorithm for Soil Classification Using Hyperspectral Features
6
作者 S.Prasanna Bharathi S.Srinivasan +1 位作者 G.Chamundeeswari B.Ramesh 《Computer Systems Science & Engineering》 SCIE EI 2023年第4期579-594,共16页
Now a days,Remote Sensing(RS)techniques are used for earth observation and for detection of soil types with high accuracy and better reliability.This technique provides perspective view of spatial resolution and aids ... Now a days,Remote Sensing(RS)techniques are used for earth observation and for detection of soil types with high accuracy and better reliability.This technique provides perspective view of spatial resolution and aids in instantaneous measurement of soil’s minerals and its characteristics.There are a few challenges that is present in soil classification using image enhancement such as,locating and plotting soil boundaries,slopes,hazardous areas,drainage condition,land use,vegetation etc.There are some traditional approaches which involves few drawbacks such as,manual involvement which results in inaccuracy due to human interference,time consuming,inconsistent prediction etc.To overcome these draw backs and to improve the predictive analysis of soil characteristics,we propose a Hybrid Deep Learning improved BAT optimization algorithm(HDIB)for soil classification using remote sensing hyperspectral features.In HDIB,we propose a spontaneous BAT optimization algorithm for feature extraction of both spectral-spatial features by choosing pure pixels from the Hyper Spectral(HS)image.Spectral-spatial vector as training illustrations is attained by merging spatial and spectral vector by means of priority stacking methodology.Then,a recurring Deep Learning(DL)Neural Network(NN)is used for classifying the HS images,considering the datasets of Pavia University,Salinas and Tamil Nadu Hill Scene,which in turn improves the reliability of classification.Finally,the performance of the proposed HDIB based soil classifier is compared and analyzed with existing methodologies like Single Layer Perceptron(SLP),Convolutional Neural Networks(CNN)and Deep Metric Learning(DML)and it shows an improved classification accuracy of 99.87%,98.34%and 99.9%for Tamil Nadu Hills dataset,Pavia University and Salinas scene datasets respectively. 展开更多
关键词 HDIB bat optimization algorithm recurrent deep learning neural network convolutional neural network single layer perceptron hyperspectral images deep metric learning
下载PDF
Deep Learning Applied to Computational Mechanics:A Comprehensive Review,State of the Art,and the Classics 被引量:1
7
作者 Loc Vu-Quoc Alexander Humer 《Computer Modeling in Engineering & Sciences》 SCIE EI 2023年第11期1069-1343,共275页
Three recent breakthroughs due to AI in arts and science serve as motivation:An award winning digital image,protein folding,fast matrix multiplication.Many recent developments in artificial neural networks,particularl... Three recent breakthroughs due to AI in arts and science serve as motivation:An award winning digital image,protein folding,fast matrix multiplication.Many recent developments in artificial neural networks,particularly deep learning(DL),applied and relevant to computational mechanics(solid,fluids,finite-element technology)are reviewed in detail.Both hybrid and pure machine learning(ML)methods are discussed.Hybrid methods combine traditional PDE discretizations with ML methods either(1)to help model complex nonlinear constitutive relations,(2)to nonlinearly reduce the model order for efficient simulation(turbulence),or(3)to accelerate the simulation by predicting certain components in the traditional integration methods.Here,methods(1)and(2)relied on Long-Short-Term Memory(LSTM)architecture,with method(3)relying on convolutional neural networks.Pure ML methods to solve(nonlinear)PDEs are represented by Physics-Informed Neural network(PINN)methods,which could be combined with attention mechanism to address discontinuous solutions.Both LSTM and attention architectures,together with modern and generalized classic optimizers to include stochasticity for DL networks,are extensively reviewed.Kernel machines,including Gaussian processes,are provided to sufficient depth for more advanced works such as shallow networks with infinite width.Not only addressing experts,readers are assumed familiar with computational mechanics,but not with DL,whose concepts and applications are built up from the basics,aiming at bringing first-time learners quickly to the forefront of research.History and limitations of AI are recounted and discussed,with particular attention at pointing out misstatements or misconceptions of the classics,even in well-known references.Positioning and pointing control of a large-deformable beam is given as an example. 展开更多
关键词 deep learning breakthroughs network architectures backpropagation stochastic optimization methods from classic to modern recurrent neural networks long short-term memory gated recurrent unit attention transformer kernel machines Gaussian processes libraries Physics-Informed Neural networks state-of-the-art history limitations challenges Applications to computational mechanics Finite-element matrix integration improved Gauss quadrature Multiscale geomechanics fluid-filled porous media Fluid mechanics turbulence proper orthogonal decomposition Nonlinear-manifold model-order reduction autoencoder hyper-reduction using gappy data control of large deformable beam
下载PDF
An adaptive machine learning-based optimization method in the aerodynamic analysis of a finite wing under various cruise conditions
8
作者 Zilan Zhang Yu Ao +1 位作者 Shaofan Li Grace X.Gu 《Theoretical & Applied Mechanics Letters》 CAS CSCD 2024年第1期27-34,共8页
Conventional wing aerodynamic optimization processes can be time-consuming and imprecise due to the complexity of versatile flight missions.Plenty of existing literature has considered two-dimensional infinite airfoil... Conventional wing aerodynamic optimization processes can be time-consuming and imprecise due to the complexity of versatile flight missions.Plenty of existing literature has considered two-dimensional infinite airfoil optimization,while three-dimensional finite wing optimizations are subject to limited study because of high computational costs.Here we create an adaptive optimization methodology built upon digitized wing shape deformation and deep learning algorithms,which enable the rapid formulation of finite wing designs for specific aerodynamic performance demands under different cruise conditions.This methodology unfolds in three stages:radial basis function interpolated wing generation,collection of inputs from computational fluid dynamics simulations,and deep neural network that constructs the surrogate model for the optimal wing configuration.It has been demonstrated that the proposed methodology can significantly reduce the computational cost of numerical simulations.It also has the potential to optimize various aerial vehicles undergoing different mission environments,loading conditions,and safety requirements. 展开更多
关键词 Aerodynamic optimization Computational fluid dynamics Radial basis function Finite wing deep learning neural network
下载PDF
A Self-Learning Data-Driven Development of Failure Criteria of Unknown Anisotropic Ductile Materials with Deep Learning Neural Network
9
作者 Kyungsuk Jang Gun Jin Yun 《Computers, Materials & Continua》 SCIE EI 2021年第2期1091-1120,共30页
This paper first proposes a new self-learning data-driven methodology that can develop the failure criteria of unknown anisotropic ductile materials from the minimal number of experimental tests.Establishing failure c... This paper first proposes a new self-learning data-driven methodology that can develop the failure criteria of unknown anisotropic ductile materials from the minimal number of experimental tests.Establishing failure criteria of anisotropic ductile materials requires time-consuming tests and manual data evaluation.The proposed method can overcome such practical challenges.The methodology is formalized by combining four ideas:1)The deep learning neural network(DLNN)-based material constitutive model,2)Self-learning inverse finite element(SELIFE)simulation,3)Algorithmic identification of failure points from the selflearned stress-strain curves and 4)Derivation of the failure criteria through symbolic regression of the genetic programming.Stress update and the algorithmic tangent operator were formulated in terms of DLNN parameters for nonlinear finite element analysis.Then,the SELIFE simulation algorithm gradually makes the DLNN model learn highly complex multi-axial stress and strain relationships,being guided by the experimental boundary measurements.Following the failure point identification,a self-learning data-driven failure criteria are eventually developed with the help of a reliable symbolic regression algorithm.The methodology and the self-learning data-driven failure criteria were verified by comparing with a reference failure criteria and simulating with different materials orientations,respectively. 展开更多
关键词 Data-driven modeling deep learning neural networks genetic programming anisotropic failure criterion
下载PDF
General seismic wave and phase detection software driven by deep learning 被引量:1
10
作者 Ming Zhao Jiahui Ma +1 位作者 Hao Chang Shi Chen 《Earthquake Research Advances》 CSCD 2021年第3期38-45,共8页
We developed an automatic seismic wave and phase detection software based on PhaseNet,an efficient and highly generalized deep learning neural network for P-and S-wave phase picking.The software organically combines m... We developed an automatic seismic wave and phase detection software based on PhaseNet,an efficient and highly generalized deep learning neural network for P-and S-wave phase picking.The software organically combines multiple modules including application terminal interface,docker container,data visualization,SSH protocol data transmission and other auxiliary modules.Characterized by a series of technologically powerful functions,the software is highly convenient for all users.To obtain the P-and S-wave picks,one only needs to prepare threecomponent seismic data as input and customize some parameters in the interface.In particular,the software can automatically identify complex waveforms(i.e.continuous or truncated waves)and support multiple types of input data such as SAC,MSEED,NumPy array,etc.A test on the dataset of the Wenchuan aftershocks shows the generalization ability and detection accuracy of the software.The software is expected to increase the efficiency and subjectivity in the manual processing of large amounts of seismic data,thereby providing convenience to regional network monitoring staffs and researchers in the study of Earth's interior. 展开更多
关键词 deep learning neural network Seismic phase detection Docker container
下载PDF
An Efficient Deep Learning-based Content-based Image Retrieval Framework 被引量:1
11
作者 M.Sivakumar N.M.Saravana Kumar N.Karthikeyan 《Computer Systems Science & Engineering》 SCIE EI 2022年第11期683-700,共18页
The use of massive image databases has increased drastically over the few years due to evolution of multimedia technology.Image retrieval has become one of the vital tools in image processing applications.Content-Base... The use of massive image databases has increased drastically over the few years due to evolution of multimedia technology.Image retrieval has become one of the vital tools in image processing applications.Content-Based Image Retrieval(CBIR)has been widely used in varied applications.But,the results produced by the usage of a single image feature are not satisfactory.So,multiple image features are used very often for attaining better results.But,fast and effective searching for relevant images from a database becomes a challenging task.In the previous existing system,the CBIR has used the combined feature extraction technique using color auto-correlogram,Rotation-Invariant Uniform Local Binary Patterns(RULBP)and local energy.However,the existing system does not provide significant results in terms of recall and precision.Also,the computational complexity is higher for the existing CBIR systems.In order to handle the above mentioned issues,the Gray Level Co-occurrence Matrix(GLCM)with Deep Learning based Enhanced Convolution Neural Network(DLECNN)is proposed in this work.The proposed system framework includes noise reduction using histogram equalization,feature extraction using GLCM,similarity matching computation using Hierarchal and Fuzzy c-Means(HFCM)algorithm and the image retrieval using DLECNN algorithm.The histogram equalization has been used for computing the image enhancement.This enhanced image has a uniform histogram.Then,the GLCM method has been used to extract the features such as shape,texture,colour,annotations and keywords.The HFCM similarity measure is used for computing the query image vector's similarity index with every database images.For enhancing the performance of this image retrieval approach,the DLECNN algorithm is proposed to retrieve more accurate features of the image.The proposed GLCM+DLECNN algorithm provides better results associated with high accuracy,precision,recall,f-measure and lesser complexity.From the experimental results,it is clearly observed that the proposed system provides efficient image retrieval for the given query image. 展开更多
关键词 Content based image retrieval(CBIR) improved gray level cooccurrence matrix(GLCM) hierarchal and fuzzy C-means(HFCM)algorithm deep learning based enhanced convolution neural network(DLECNN)
下载PDF
HDLIDP: A Hybrid Deep Learning Intrusion Detection and Prevention Framework
12
作者 Magdy M.Fadel Sally M.El-Ghamrawy +2 位作者 Amr M.T.Ali-Eldin Mohammed K.Hassan Ali I.El-Desoky 《Computers, Materials & Continua》 SCIE EI 2022年第11期2293-2312,共20页
Distributed denial-of-service(DDoS)attacks are designed to interrupt network services such as email servers and webpages in traditional computer networks.Furthermore,the enormous number of connected devices makes it d... Distributed denial-of-service(DDoS)attacks are designed to interrupt network services such as email servers and webpages in traditional computer networks.Furthermore,the enormous number of connected devices makes it difficult to operate such a network effectively.Software defined networks(SDN)are networks that are managed through a centralized control system,according to researchers.This controller is the brain of any SDN,composing the forwarding table of all data plane network switches.Despite the advantages of SDN controllers,DDoS attacks are easier to perpetrate than on traditional networks.Because the controller is a single point of failure,if it fails,the entire network will fail.This paper offers a Hybrid Deep Learning Intrusion Detection and Prevention(HDLIDP)framework,which blends signature-based and deep learning neural networks to detect and prevent intrusions.This framework improves detection accuracy while addressing all of the aforementioned problems.To validate the framework,experiments are done on both traditional and SDN datasets;the findings demonstrate a significant improvement in classification accuracy. 展开更多
关键词 Software defined networks(SDN) distributed denial of service attack(DDoS) signature-based detection whale optimization algorism(WOA) deep learning neural network classifier
下载PDF
基于水动力载荷混合数据集的高精度神经网络代理模型构建
13
作者 敖愈 李云波 +1 位作者 李少凡 龚家烨 《哈尔滨工程大学学报(英文版)》 CSCD 2024年第1期49-63,共15页
In this work,we constructed a neural network proxy model(NNPM)to estimate the hydrodynamic resistance in the ship hull structure design process,which is based on the hydrodynamic load data obtained from both the poten... In this work,we constructed a neural network proxy model(NNPM)to estimate the hydrodynamic resistance in the ship hull structure design process,which is based on the hydrodynamic load data obtained from both the potential flow method(PFM)and the viscous flow method(VFM).Here the PFM dataset is applied for the tuning,pre-training,and the VFM dataset is applied for the fine-training.By adopting the PFM and VFM datasets simultaneously,we aim to construct an NNPM to achieve the high-accuracy prediction on hydrodynamic load on ship hull structures exerted from the viscous flow,while ensuring a moderate data-acquiring workload.The high accuracy prediction on hydrodynamic loads and the relatively low dataset establishment cost of the NNPM developed demonstrated the effectiveness and feasibility of hybrid dataset based NNPM achieving a high precision prediction of hydrodynamic loads on ship hull structures.The successful construction of the high precision hydrodynamic prediction NNPM advances the artificial intelligence-assisted design(AIAD)technology for various marine structures. 展开更多
关键词 deep learning neural network Hybrid dataset Proxy model Ship hull design Machine learning
下载PDF
Forecasting solar power generation using evolutionary mating algorithm-deep neural networks
14
作者 Mohd Herwan Sulaiman Zuriani Mustaffa 《Energy and AI》 EI 2024年第2期346-362,共17页
This paper proposes an integration of recent metaheuristic algorithm namely Evolutionary Mating Algorithm (EMA) in optimizing the weights and biases of deep neural networks (DNN) for forecasting the solar power genera... This paper proposes an integration of recent metaheuristic algorithm namely Evolutionary Mating Algorithm (EMA) in optimizing the weights and biases of deep neural networks (DNN) for forecasting the solar power generation. The study employs a Feed Forward Neural Network (FFNN) to forecast AC power output using real solar power plant measurements spanning a 34-day period, recorded at 15-minute intervals. The intricate nonlinear relationship between solar irradiation, ambient temperature, and module temperature is captured for accurate prediction. Additionally, the paper conducts a comprehensive comparison with established algorithms, including Differential Evolution (DE-DNN), Barnacles Mating Optimizer (BMO-DNN), Particle Swarm Optimization (PSO-DNN), Harmony Search Algorithm (HSA-DNN), DNN with Adaptive Moment Estimation optimizer (ADAM) and Nonlinear AutoRegressive with eXogenous inputs (NARX). The experimental results distinctly highlight the exceptional performance of EMA-DNN by attaining the lowest Root Mean Squared Error (RMSE) during testing. This contribution not only advances solar power forecasting methodologies but also underscores the potential of merging metaheuristic algorithms with contemporary neural networks for improved accuracy and reliability. 展开更多
关键词 deep learning neural networks Evolutionary mating algorithm Feed forward neural networks Metaheuristic Optimizers Solar PV
下载PDF
Integrated framework for seismic fragility assessment of cable-stayed bridges using deep learning neural networks
15
作者 PANG YuTao YIN PengCheng +1 位作者 WANG JianGuo WU Li 《Science China(Technological Sciences)》 SCIE EI CAS CSCD 2023年第2期406-416,共11页
The increasing intensity of strong earthquakes has a large impact on the seismic safety of bridges worldwide.As the key component in the transportation network,the cable-stayed bridge should cope with the increasing f... The increasing intensity of strong earthquakes has a large impact on the seismic safety of bridges worldwide.As the key component in the transportation network,the cable-stayed bridge should cope with the increasing future hazards to improve seismic safety.Seismic fragility analysis can assist the resilience assessment under different levels of seismic intensity.However,such an analysis is computationally intensive,especially when considering various random factors.The present paper implemented the deep learning neural networks that are integrated into the performance-based earthquake engineering framework to predict fragility functions and associated resilience index of cable-stayed bridges under seismic hazards to improve the computational efficiency while having sufficient accuracy.In the proposed framework,the Latin hypercube sampling was improved with additional uniformity to enhance the training process of the neural network.The well-trained neural network was then applied in a probabilistic simulation process to derive different component fragilities of the cable-stayed bridge.The estimated fragility functions were combined with the Monte Carlo simulations to predict system resilience.The proposed integrated framework in this study was demonstrated on an existing single-pylon cable-stayed bridge in China.Results reveal that this integrated framework yields accurate predictions of fragility functions for the cable-stayed bridge and has reasonable accuracy compared with the conventional methods. 展开更多
关键词 deep learning neural network cable-stayed bridge fragility analysis uniform design method seismic resilience
原文传递
Chiller faults detection and diagnosis with sensor network and adaptive 1D CNN 被引量:3
16
作者 Ke Yan Xiaokang Zhou 《Digital Communications and Networks》 SCIE CSCD 2022年第4期531-539,共9页
Computer-empowered detection of possible faults for Heating,Ventilation and Air-Conditioning(HVAC)subsystems,e.g.,chillers,is one of the most important applications in Artificial Intelligence(AI)integrated Internet of... Computer-empowered detection of possible faults for Heating,Ventilation and Air-Conditioning(HVAC)subsystems,e.g.,chillers,is one of the most important applications in Artificial Intelligence(AI)integrated Internet of Things(IoT).The cyber-physical system greatly enhances the safety and security of the working facilities,reducing time,saving energy and protecting humans’health.Under the current trends of smart building design and energy management optimization,Automated Fault Detection and Diagnosis(AFDD)of chillers integrated with IoT is highly demanded.Recent studies show that standard machine learning techniques,such as Principal Component Analysis(PCA),Support Vector Machine(SVM)and tree-structure-based algorithms,are useful in capturing various chiller faults with high accuracy rates.With the fast development of deep learning technology,Convolutional Neural Networks(CNNs)have been widely and successfully applied to various fields.However,for chiller AFDD,few existing works are adopting CNN and its extensions in the feature extraction and classification processes.In this study,we propose to perform chiller FDD using a CNN-based approach.The proposed approach has two distinct advantages over existing machine learning-based chiller AFDD methods.First,the CNN-based approach does not require the feature selection/extraction process.Since CNN is reputable with its feature extraction capability,the feature extraction and classification processes are merged,leading to a more neat AFDD framework compared to traditional approaches.Second,the classification accuracy is significantly improved compared to traditional methods using the CNN-based approach. 展开更多
关键词 CHILLER Fault detection and diagnosis deep learning neural network Long short term memory Recurrent neural network Gated recurrent unit
下载PDF
Solar image deconvolution by generative adversarial network 被引量:2
17
作者 Long Xu Wen-Qing Sun +1 位作者 Yi-Hua Yan Wei-Qiang Zhang 《Research in Astronomy and Astrophysics》 SCIE CAS CSCD 2020年第11期182-190,共9页
With aperture synthesis(AS)technique,a number of small antennas can be assembled to form a large telescope whose spatial resolution is determined by the distance of two farthest antennas instead of the diameter of a s... With aperture synthesis(AS)technique,a number of small antennas can be assembled to form a large telescope whose spatial resolution is determined by the distance of two farthest antennas instead of the diameter of a single-dish antenna.In contrast from a direct imaging system,an AS telescope captures the Fourier coefficients of a spatial object,and then implement inverse Fourier transform to reconstruct the spatial image.Due to the limited number of antennas,the Fourier coefficients are extremely sparse in practice,resulting in a very blurry image.To remove/reduce blur,“CLEAN”deconvolution has been widely used in the literature.However,it was initially designed for a point source.For an extended source,like the Sun,its efficiency is unsatisfactory.In this study,a deep neural network,referring to Generative Adversarial Network(GAN),is proposed for solar image deconvolution.The experimental results demonstrate that the proposed model is markedly better than traditional CLEAN on solar images.The main purpose of this work is visual inspection instead of quantitative scientific computation.We believe that this will also help scientists to better understand solar phenomena with high quality images. 展开更多
关键词 deep learning(DL)generative adversarial network(GAN)solar radio astronomy
下载PDF
Aerial multi-spectral AI-based detection system for unexploded ordnance 被引量:1
18
作者 Seungwan Cho Jungmok Ma Oleg A.Yakimenko 《Defence Technology(防务技术)》 SCIE EI CAS CSCD 2023年第9期24-37,共14页
Unexploded ordnance(UXO)poses a threat to soldiers operating in mission areas,but current UXO detection systems do not necessarily provide the required safety and efficiency to protect soldiers from this hazard.Recent... Unexploded ordnance(UXO)poses a threat to soldiers operating in mission areas,but current UXO detection systems do not necessarily provide the required safety and efficiency to protect soldiers from this hazard.Recent technological advancements in artificial intelligence(AI)and small unmanned aerial systems(sUAS)present an opportunity to explore a novel concept for UXO detection.The new UXO detection system proposed in this study takes advantage of employing an AI-trained multi-spectral(MS)sensor on sUAS.This paper explores feasibility of AI-based UXO detection using sUAS equipped with a single(visible)spectrum(SS)or MS digital electro-optical(EO)sensor.Specifically,it describes the design of the Deep Learning Convolutional Neural Network for UXO detection,the development of an AI-based algorithm for reliable UXO detection,and also provides a comparison of performance of the proposed system based on SS and MS sensor imagery. 展开更多
关键词 Unexploded ordnance(UXO) Multispectral imaging Small unmanned aerial systems(sUAS) Object detection deep learning convolutional neural network(DLCNN)
下载PDF
Optimization Scheme of Trusted Task Offloading in IIoT Scenario Based on DQN 被引量:2
19
作者 Xiaojuan Wang Zikui Lu +3 位作者 Siyuan Sun Jingyue Wang Luona Song Merveille Nicolas 《Computers, Materials & Continua》 SCIE EI 2023年第1期2055-2071,共17页
With the development of the Industrial Internet of Things(IIoT),end devices(EDs)are equipped with more functions to capture information.Therefore,a large amount of data is generated at the edge of the network and need... With the development of the Industrial Internet of Things(IIoT),end devices(EDs)are equipped with more functions to capture information.Therefore,a large amount of data is generated at the edge of the network and needs to be processed.However,no matter whether these computing tasks are offloaded to traditional central clusters or mobile edge computing(MEC)devices,the data is short of security and may be changed during transmission.In view of this challenge,this paper proposes a trusted task offloading optimization scheme that can offer low latency and high bandwidth services for IIoT with data security.Blockchain technology is adopted to ensure data consistency.Meanwhile,to reduce the impact of low throughput of blockchain on task offloading performance,we design the processes of consensus and offloading as a Markov decision process(MDP)by defining states,actions,and rewards.Deep reinforcement learning(DRL)algorithm is introduced to dynamically select offloading actions.To accelerate the optimization,we design a novel reward function for the DRL algorithm according to the scale and computational complexity of the task.Experiments demonstrate that compared with methods without optimization,our mechanism performs better when it comes to the number of task offloading and throughput of blockchain. 展开更多
关键词 Task offloading blockchain industrial internet of things(IIoT) deep reinforcement learning(DRL)network mobile-edge computing(MEC)
下载PDF
Modulation recognition network of multi-scale analysis with deep threshold noise elimination
20
作者 Xiang LI Yibing LI +1 位作者 Chunrui TANG Yingsong LI 《Frontiers of Information Technology & Electronic Engineering》 SCIE EI CSCD 2023年第5期742-758,共17页
To improve the accuracy of modulated signal recognition in variable environments and reduce the impact of factors such as lack of prior knowledge on recognition results,researchers have gradually adopted deep learning... To improve the accuracy of modulated signal recognition in variable environments and reduce the impact of factors such as lack of prior knowledge on recognition results,researchers have gradually adopted deep learning techniques to replace traditional modulated signal processing techniques.To address the problem of low recognition accuracy of the modulated signal at low signal-to-noise ratios,we have designed a novel modulation recognition network of multi-scale analysis with deep threshold noise elimination to recognize the actually collected modulated signals under a symmetric cross-entropy function of label smoothing.The network consists of a denoising encoder with deep adaptive threshold learning and a decoder with multi-scale feature fusion.The two modules are skip-connected to work together to improve the robustness of the overall network.Experimental results show that this method has better recognition accuracy at low signal-to-noise ratios than previous methods.The network demonstrates a flexible self-learning capability for different noise thresholds and the effectiveness of the designed feature fusion module in multi-scale feature acquisition for various modulation types. 展开更多
关键词 Signal noise elimination deep adaptive threshold learning network Multi-scale feature fusion Modulation ecognition
原文传递
上一页 1 2 下一页 到第
使用帮助 返回顶部